modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-09-07 00:41:44
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
544 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-09-07 00:41:34
card
stringlengths
11
1.01M
ganghe74/distilbert-base-uncased-finetuned-emotion
ganghe74
2023-06-17T09:34:40Z
105
0
transformers
[ "transformers", "pytorch", "distilbert", "text-classification", "generated_from_trainer", "dataset:emotion", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-06-17T09:13:51Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - emotion metrics: - accuracy - f1 model-index: - name: distilbert-base-uncased-finetuned-emotion results: - task: name: Text Classification type: text-classification dataset: name: emotion type: emotion config: split split: validation args: split metrics: - name: Accuracy type: accuracy value: 0.9225 - name: F1 type: f1 value: 0.922469380812715 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.2170 - Accuracy: 0.9225 - F1: 0.9225 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.8057 | 1.0 | 250 | 0.3170 | 0.905 | 0.9023 | | 0.242 | 2.0 | 500 | 0.2170 | 0.9225 | 0.9225 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1 - Datasets 2.13.0 - Tokenizers 0.13.3
edu-linguistic/opt-1.3b-edu-sft
edu-linguistic
2023-06-17T09:28:57Z
0
0
null
[ "en", "dataset:yahma/alpaca-cleaned", "dataset:Nebulous/gpt4all_pruned", "region:us" ]
null
2023-06-15T14:16:11Z
--- datasets: - yahma/alpaca-cleaned - Nebulous/gpt4all_pruned language: - en --- ## Inference Example: ```python from peft import PeftModel, PeftConfig from transformers import AutoModelForCausalLM, AutoTokenizer peft_model_id = "edu-linguistic/opt-1.3b-edu-sft" model_name = 'facebook/opt-1.3b' config = PeftConfig.from_pretrained(peft_model_id) model = AutoModelForCausalLM.from_pretrained(model_name) model = PeftModel.from_pretrained(model, peft_model_id) tokenizer = AutoTokenizer.from_pretrained(model_name) question = "<|prompter|> Consider the following function: f(x1, x2) = ln(x1). This function is…" question = tokenizer.encode(question, return_tensors='pt') generation_kwargs = { "do_sample": True, "top_k": 0, "top_p": 0.9, "bos_token_id": tokenizer.bos_token_id, "pad_token_id": tokenizer.pad_token_id, "eos_token_id": tokenizer.eos_token_id, "num_return_sequences": 1, "min_new_tokens": 10, "max_new_tokens": 512, } response = model.generate(input_ids=question, **generation_kwargs) response = tokenizer.decode(response[0], skip_special_tokens=False, clean_up_tokenization_spaces=False ) print(response) ```
benol/Roma_Pyatifan
benol
2023-06-17T09:10:31Z
0
0
null
[ "ru", "en", "arxiv:1910.09700", "license:unknown", "region:us" ]
null
2023-06-17T08:58:04Z
--- license: unknown language: - ru - en --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Data Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
TheBloke/robin-65B-v2-GGML
TheBloke
2023-06-17T08:01:48Z
0
17
null
[ "license:other", "region:us" ]
null
2023-06-16T21:59:56Z
--- inference: false license: other --- <!-- header start --> <div style="width: 100%;"> <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <div style="display: flex; justify-content: space-between; width: 100%;"> <div style="display: flex; flex-direction: column; align-items: flex-start;"> <p><a href="https://discord.gg/Jq4vkcDakD">Chat & support: my new Discord server</a></p> </div> <div style="display: flex; flex-direction: column; align-items: flex-end;"> <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p> </div> </div> <!-- header end --> # OptimalScale's Robin 65B v2 GGML These files are GGML format model files for [OptimalScale's Robin 65B v2](https://huggingface.co/OptimalScale/robin-65b-v2-delta). GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as: * [text-generation-webui](https://github.com/oobabooga/text-generation-webui) * [KoboldCpp](https://github.com/LostRuins/koboldcpp) * [ParisNeo/GPT4All-UI](https://github.com/ParisNeo/gpt4all-ui) * [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) * [ctransformers](https://github.com/marella/ctransformers) ## Repositories available * [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/robin-65B-v2-GPTQ) * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/robin-65B-v2-GGML) * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/robin-65b-v2-fp16) ## Prompt template ``` A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions ###Human: prompt ###Assistant: ``` <!-- compatibility_ggml start --> ## Compatibility ### Original llama.cpp quant methods: `q4_0, q4_1, q5_0, q5_1, q8_0` I have quantized these 'original' quantisation methods using an older version of llama.cpp so that they remain compatible with llama.cpp as of May 19th, commit `2d5db48`. These are guaranteed to be compatbile with any UIs, tools and libraries released since late May. ### New k-quant methods: `q2_K, q3_K_S, q3_K_M, q3_K_L, q4_K_S, q4_K_M, q5_K_S, q6_K` These new quantisation methods are compatible with llama.cpp as of June 6th, commit `2d43387`. They are now also compatible with recent releases of text-generation-webui, KoboldCpp, llama-cpp-python and ctransformers. Other tools and libraries may or may not be compatible - check their documentation if in doubt. ## Explanation of the new k-quant methods The new methods available are: * GGML_TYPE_Q2_K - "type-1" 2-bit quantization in super-blocks containing 16 blocks, each block having 16 weight. Block scales and mins are quantized with 4 bits. This ends up effectively using 2.5625 bits per weight (bpw) * GGML_TYPE_Q3_K - "type-0" 3-bit quantization in super-blocks containing 16 blocks, each block having 16 weights. Scales are quantized with 6 bits. This end up using 3.4375 bpw. * GGML_TYPE_Q4_K - "type-1" 4-bit quantization in super-blocks containing 8 blocks, each block having 32 weights. Scales and mins are quantized with 6 bits. This ends up using 4.5 bpw. * GGML_TYPE_Q5_K - "type-1" 5-bit quantization. Same super-block structure as GGML_TYPE_Q4_K resulting in 5.5 bpw * GGML_TYPE_Q6_K - "type-0" 6-bit quantization. Super-blocks with 16 blocks, each block having 16 weights. Scales are quantized with 8 bits. This ends up using 6.5625 bpw * GGML_TYPE_Q8_K - "type-0" 8-bit quantization. Only used for quantizing intermediate results. The difference to the existing Q8_0 is that the block size is 256. All 2-6 bit dot products are implemented for this quantization type. Refer to the Provided Files table below to see what files use which methods, and how. <!-- compatibility_ggml end --> ## Provided files | Name | Quant method | Bits | Size | Max RAM required | Use case | | ---- | ---- | ---- | ---- | ---- | ----- | | robin-65b.ggmlv3.q2_K.bin | q2_K | 2 | 27.45 GB | 29.95 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.vw and feed_forward.w2 tensors, GGML_TYPE_Q2_K for the other tensors. | | robin-65b.ggmlv3.q3_K_L.bin | q3_K_L | 3 | 34.65 GB | 37.15 GB | New k-quant method. Uses GGML_TYPE_Q5_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | robin-65b.ggmlv3.q3_K_M.bin | q3_K_M | 3 | 31.50 GB | 34.00 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | robin-65b.ggmlv3.q3_K_S.bin | q3_K_S | 3 | 28.16 GB | 30.66 GB | New k-quant method. Uses GGML_TYPE_Q3_K for all tensors | | robin-65b.ggmlv3.q4_0.bin | q4_0 | 4 | 36.73 GB | 39.23 GB | Original llama.cpp quant method, 4-bit. | | robin-65b.ggmlv3.q4_1.bin | q4_1 | 4 | 40.81 GB | 43.31 GB | Original llama.cpp quant method, 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models. | | robin-65b.ggmlv3.q4_K_M.bin | q4_K_M | 4 | 39.35 GB | 41.85 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q4_K | | robin-65b.ggmlv3.q4_K_S.bin | q4_K_S | 4 | 36.80 GB | 39.30 GB | New k-quant method. Uses GGML_TYPE_Q4_K for all tensors | | robin-65b.ggmlv3.q5_0.bin | q5_0 | 5 | 44.89 GB | 47.39 GB | Original llama.cpp quant method, 5-bit. Higher accuracy, higher resource usage and slower inference. | | robin-65b.ggmlv3.q5_1.bin | q5_1 | 5 | 48.97 GB | 51.47 GB | Original llama.cpp quant method, 5-bit. Even higher accuracy, resource usage and slower inference. | | robin-65b.ggmlv3.q5_K_M.bin | q5_K_M | 5 | 46.24 GB | 48.74 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q5_K | | robin-65b.ggmlv3.q5_K_S.bin | q5_K_S | 5 | 44.92 GB | 47.42 GB | New k-quant method. Uses GGML_TYPE_Q5_K for all tensors | | robin-65b.ggmlv3.q6_K.bin | q6_K | 6 | 53.56 GB | 56.06 GB | New k-quant method. Uses GGML_TYPE_Q8_K - 6-bit quantization - for all tensors | | robin-65b.ggmlv3.q8_0.bin | q8_0 | 8 | 69.370 GB | 71.87 GB | Original llama.cpp quant method, 8-bit. Almost indistinguishable from float16. High resource use and slow. Not recommended for most users. | **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead. ### q6_K and q8_0 files require expansion from archive **Note:** HF does not support uploading files larger than 50GB. Therefore I have uploaded the q6_K and q8_0 files as multi-part ZIP files. They are not compressed, it is just storing the .bin file in two parts. ### q6_K Please download: * `robin-65b.ggmlv3.q6_K.zip` * `robin-65b.ggmlv3.q6_K.z01` ### q8_0 Please download: * `robin-65b.ggmlv3.q8_0.zip` * `robin-65b.ggmlv3.q8_0.z01` Then extract the .zip archive. This will will expand both parts automatically. On Linux I found I had to use `7zip` - the basic `unzip` tool did not work. Example: ``` sudo apt update -y && sudo apt install 7zip 7zz x robin-65b.ggmlv3.q6_K.zip` ``` Once the `.bin` is extracted you can delete the `.zip` and `.z01` files ## How to run in `llama.cpp` I use the following command line; adjust for your tastes and needs: ``` ./main -t 10 -ngl 32 -m robin-65b.ggmlv3.q5_0.bin --color -c 2048 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions\n###Human: write a story about llamas\n###Assistant: ``` Change `-t 10` to the number of physical CPU cores you have. For example if your system has 8 cores/16 threads, use `-t 8`. Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration. If you want to have a chat-style conversation, replace the `-p <PROMPT>` argument with `-i -ins` ## How to run in `text-generation-webui` Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md). <!-- footer start --> ## Discord For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/Jq4vkcDakD) ## Thanks, and how to contribute. Thanks to the [chirper.ai](https://chirper.ai) team! I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. * Patreon: https://patreon.com/TheBlokeAI * Ko-Fi: https://ko-fi.com/TheBlokeAI **Special thanks to**: Luke from CarbonQuill, Aemon Algiz, Dmitriy Samsonov. **Patreon special mentions**: vamX, K, Jonathan Leane, Lone Striker, Sean Connelly, Chris McCloskey, WelcomeToTheClub, Nikolai Manek, John Detwiler, Kalila, David Flickinger, Fen Risland, subjectnull, Johann-Peter Hartmann, Talal Aujan, John Villwock, senxiiz, Khalefa Al-Ahmad, Kevin Schuppel, Alps Aficionado, Derek Yates, Mano Prime, Nathan LeClaire, biorpg, trip7s trip, Asp the Wyvern, chris gileta, Iucharbius , Artur Olbinski, Ai Maven, Joseph William Delisle, Luke Pendergrass, Illia Dulskyi, Eugene Pentland, Ajan Kanaga, Willem Michiel, Space Cruiser, Pyrater, Preetika Verma, Junyu Yang, Oscar Rangel, Spiking Neurons AB, Pierre Kircher, webtim, Cory Kujawski, terasurfer , Trenton Dambrowitz, Gabriel Puliatti, Imad Khwaja, Luke. Thank you to all my generous patrons and donaters! <!-- footer end --> # Original model card: OptimalScale's Robin 65B v2 No model card provided in source repository.
musabg/mt5-xl-tr-summarization
musabg
2023-06-17T07:25:20Z
5
1
transformers
[ "transformers", "pytorch", "tensorboard", "mt5", "text2text-generation", "generated_from_trainer", "tr", "dataset:musabg/wikipedia-tr-summarization", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2023-06-08T16:24:59Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - musabg/wikipedia-tr-summarization metrics: - rouge model-index: - name: mt5-xl-tr-summarization results: - task: name: Summarization type: summarization dataset: name: musabg/wikipedia-tr-summarization type: musabg/wikipedia-tr-summarization split: validation metrics: - name: Rouge1 type: rouge value: 56.4468 language: - tr --- # mT5-Xl Turkish Summarization This model is a fine-tuned version of [google/mt5-xl](https://huggingface.co/google/mt5-xl) on the musabg/wikipedia-tr-summarization dataset. This can be used with HF summarization pipeline. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 8 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1.0 ### Eval results It achieves the following results on the evaluation set: - Loss: 0.5676 - Rouge1: 56.4468 - Rouge2: 41.3258 - Rougel: 48.1909 - Rougelsum: 48.4284 - Gen Len: 75.9265 ### Training results ### Framework versions - Transformers 4.31.0.dev0 - Pytorch 1.13.1 - Datasets 2.12.0 - Tokenizers 0.13.3
nolanaatama/spngbbsqrpnts1kpchs99kstpslhmnn
nolanaatama
2023-06-17T07:14:41Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-17T07:05:05Z
--- license: creativeml-openrail-m ---
kjiwon1222/my_awesome_eli5_clm-model
kjiwon1222
2023-06-17T06:54:34Z
217
0
transformers
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2023-06-17T06:32:49Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: my_awesome_eli5_clm-model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_awesome_eli5_clm-model This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 3.7506 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.8621 | 1.0 | 1137 | 3.7690 | | 3.7782 | 2.0 | 2274 | 3.7533 | | 3.7245 | 3.0 | 3411 | 3.7506 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3
aga3134/poca-SoccerTwos
aga3134
2023-06-17T06:48:55Z
0
0
ml-agents
[ "ml-agents", "tensorboard", "SoccerTwos", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-SoccerTwos", "region:us" ]
reinforcement-learning
2023-06-17T06:48:14Z
--- library_name: ml-agents tags: - SoccerTwos - deep-reinforcement-learning - reinforcement-learning - ML-Agents-SoccerTwos --- # **poca** Agent playing **SoccerTwos** This is a trained model of a **poca** agent playing **SoccerTwos** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: aga3134/poca-SoccerTwos 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
octipuw/policy_gradient-cartpole-v1
octipuw
2023-06-17T06:40:28Z
0
0
null
[ "CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2023-06-17T05:35:35Z
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: policy_gradient-cartpole-v1 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 166.70 +/- 18.03 name: mean_reward verified: false --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
vedu/bart-large-perturbed
vedu
2023-06-17T06:21:37Z
103
0
transformers
[ "transformers", "pytorch", "jax", "rust", "bart", "feature-extraction", "en", "arxiv:1910.13461", "license:apache-2.0", "endpoints_compatible", "region:us" ]
feature-extraction
2023-06-16T20:22:31Z
--- license: apache-2.0 language: en --- # BART (large-sized model) ## Model description BART is a transformer encoder-decoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. BART is pre-trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text. BART is particularly effective when fine-tuned for text generation (e.g. summarization, translation) but also works well for comprehension tasks (e.g. text classification, question answering). Weights shared here are effectively from facebook/bart-large but with added noise for BOS embedding to assist the finetuning. ## Intended uses & limitations There have been quite a few issues related to finetuning BART for text generation, and this repo implements solution discussed in [#15559](https://github.com/huggingface/transformers/issues/15559). Particularly adding some noise to pre-trained model's BOS embedding. This seems to solve the problem of endless BOS generation for a finetuned BART model. You can use the raw model for text infilling. However, the model is mostly meant to be fine-tuned on a supervised dataset. See the [model hub](https://huggingface.co/models?search=bart) to look for fine-tuned versions on a task that interests you. ### How to use Here is how to use this model in PyTorch: ```python from transformers import BartTokenizer, BartModel tokenizer = BartTokenizer.from_pretrained('vedu/bart-large-perturbed') model = BartModel.from_pretrained('vedu/bart-large-perturbed') inputs = tokenizer("Hello, my dog is cute", return_tensors="pt") outputs = model(**inputs) last_hidden_states = outputs.last_hidden_state ``` ### BibTeX entry and citation info ```bibtex @article{DBLP:journals/corr/abs-1910-13461, author = {Mike Lewis and Yinhan Liu and Naman Goyal and Marjan Ghazvininejad and Abdelrahman Mohamed and Omer Levy and Veselin Stoyanov and Luke Zettlemoyer}, title = {{BART:} Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension}, journal = {CoRR}, volume = {abs/1910.13461}, year = {2019}, url = {http://arxiv.org/abs/1910.13461}, eprinttype = {arXiv}, eprint = {1910.13461}, timestamp = {Thu, 31 Oct 2019 14:02:26 +0100}, biburl = {https://dblp.org/rec/journals/corr/abs-1910-13461.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ```
pigeon01/sungju-finetuned-zh-to-ko1
pigeon01
2023-06-17T05:47:05Z
228
0
transformers
[ "transformers", "pytorch", "tensorboard", "m2m_100", "text2text-generation", "translation", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
translation
2023-06-17T05:12:16Z
--- license: mit tags: - translation - generated_from_trainer metrics: - bleu model-index: - name: sungju-finetuned-zh-to-ko1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # sungju-finetuned-zh-to-ko1 This model is a fine-tuned version of [alirezamsh/small100](https://huggingface.co/alirezamsh/small100) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 2.0467 - Bleu: 10.2096 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3
nolanaatama/skrmkhllvjprvc500pchsmgzb
nolanaatama
2023-06-17T05:02:20Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-17T04:59:21Z
--- license: creativeml-openrail-m ---
nolanaatama/rngrndrvcv800pchsrthysttylrsvrsn
nolanaatama
2023-06-17T04:33:50Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-05-31T04:13:59Z
--- license: creativeml-openrail-m ---
aozora-esther/aozora-esther
aozora-esther
2023-06-17T04:00:52Z
0
0
null
[ "license:bigscience-openrail-m", "region:us" ]
null
2023-06-17T04:00:52Z
--- license: bigscience-openrail-m ---
2022happy/swin-tiny-patch4-window7-224-finetuned-eurosat
2022happy
2023-06-17T03:51:48Z
245
0
transformers
[ "transformers", "pytorch", "tensorboard", "swin", "image-classification", "generated_from_trainer", "dataset:cifar10", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2023-06-15T13:46:26Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - cifar10 metrics: - accuracy model-index: - name: swin-tiny-patch4-window7-224-finetuned-eurosat results: - task: name: Image Classification type: image-classification dataset: name: cifar10 type: cifar10 config: plain_text split: train args: plain_text metrics: - name: Accuracy type: accuracy value: 0.97 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # swin-tiny-patch4-window7-224-finetuned-eurosat This model is a fine-tuned version of [microsoft/swin-tiny-patch4-window7-224](https://huggingface.co/microsoft/swin-tiny-patch4-window7-224) on the cifar10 dataset. It achieves the following results on the evaluation set: - Loss: 0.0893 - Accuracy: 0.97 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.5255 | 1.0 | 351 | 0.1262 | 0.9596 | | 0.3808 | 2.0 | 703 | 0.1031 | 0.9652 | | 0.3268 | 2.99 | 1053 | 0.0893 | 0.97 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3
mehnaazasad/bart-large-finetuned-arxiv-co-ga-old
mehnaazasad
2023-06-17T03:05:43Z
11
0
transformers
[ "transformers", "pytorch", "tensorboard", "bart", "text2text-generation", "en", "dataset:mehnaazasad/arxiv_astro_co_ga", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2023-06-02T23:10:50Z
--- license: mit datasets: - mehnaazasad/arxiv_astro_co_ga language: - en --- <span style="color:indianred">Status: As of June 8th, 2023 this model has been archived.</span> For most recent version, please visit https://huggingface.co/mehnaazasad/bart-large-finetuned-arxiv-co-ga-latest
nolanaatama/dmnslyrkmtsnybnmstyllr
nolanaatama
2023-06-17T02:31:22Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-17T02:27:58Z
--- license: creativeml-openrail-m ---
sharpbai/alpaca-lora-7b-reproduced
sharpbai
2023-06-17T02:21:47Z
0
0
null
[ "dataset:yahma/alpaca-cleaned", "license:mit", "region:us" ]
null
2023-06-16T16:24:33Z
--- license: mit datasets: - yahma/alpaca-cleaned --- This repo reproduced [tloen/alpaca-lora-7b](https://huggingface.co/tloen/alpaca-lora-7b) fit on the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) dataset. 4x H100 training for about 1h15min, details in [W&B link](https://wandb.ai/sharpbai/alpaca-lora-reproduce/runs/08ulvstd), there is a hyperparameter of val_set_size=500 4 x 4090 training for about 4h35min, details in [W&B link](https://wandb.ai/sharpbai/alpaca-lora-reproduce/runs/ws16av1u), all key hyperparameters are the same To optimize the running speed, I change these code - `load_in_8bits=False` to use 16bit finetune - comment `model = prepare_model_for_int8_training` to not turn some parameters to fp32 and turn off gradient checkpointing - for 4090 enable gradient checkpointing, add `model.gradient_checkpointing_enable()` and `model.enable_input_require_grads()` This version of the weights was trained with the following hyperparameters: - Epochs: 10 (load from best epoch) - Batch size: 128 - Cutoff length: 512 - Learning rate: 3e-4 - Lora _r_: 16 - Lora target modules: q_proj, k_proj, v_proj, o_proj That is: ``` python finetune.py \ --base_model='decapoda-research/llama-7b-hf' \ --num_epochs=10 \ --cutoff_len=512 \ --group_by_length \ --val_set_size=500 \ --output_dir='./alpaca-lora-train-H100-80G-HBM3x4-mb8' \ --lora_target_modules='[q_proj,k_proj,v_proj,o_proj]' \ --lora_r=16 \ --micro_batch_size=8 \ --train_in_8bit False ``` Instructions for running it can be found at https://github.com/tloen/alpaca-lora.
DreamerGPT/D7b-5-1
DreamerGPT
2023-06-17T01:38:49Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2023-06-17T01:20:31Z
--- license: apache-2.0 --- # D7b-5-1 [https://github.com/DreamerGPT/DreamerGPT](https://github.com/DreamerGPT/DreamerGPT)
mskani/controlnet-hands
mskani
2023-06-17T01:35:09Z
0
5
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T14:10:39Z
--- license: creativeml-openrail-m ---
zhangjian94cn/Taxi-v3
zhangjian94cn
2023-06-17T01:33:35Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-06-17T01:33:26Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: Taxi-v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.56 +/- 2.71 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="zhangjian94cn/Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
DreamerGPT/D13b-3-3
DreamerGPT
2023-06-17T01:21:55Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2023-06-17T00:58:23Z
--- license: apache-2.0 --- # D13b-3-3 [https://github.com/DreamerGPT/DreamerGPT](https://github.com/DreamerGPT/DreamerGPT)
harshseth/distilbert-base-uncased-finetuned-imdb
harshseth
2023-06-17T01:07:35Z
123
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "fill-mask", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2023-06-16T19:14:46Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: distilbert-base-uncased-finetuned-imdb results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-imdb This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.2646 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.6157 | 1.0 | 469 | 2.3844 | | 2.4501 | 2.0 | 938 | 2.2822 | | 2.377 | 3.0 | 1407 | 2.2549 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3
sheshenin/vvshsh
sheshenin
2023-06-17T00:40:05Z
32
0
diffusers
[ "diffusers", "safetensors", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2023-06-17T00:35:21Z
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### VikaSH Dreambooth model trained by sheshenin with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Sample pictures of this concept:
Ioanaaaaaaa/bert-base-uncased-with-preprocess-finetuned-emotion-5-epochs-5e-05-lr-0.1-weight_decay
Ioanaaaaaaa
2023-06-16T23:47:54Z
105
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "dataset:emotion", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-06-16T23:30:14Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - emotion metrics: - accuracy - f1 model-index: - name: bert-base-uncased-with-preprocess-finetuned-emotion-5-epochs-5e-05-lr-0.1-weight_decay results: - task: name: Text Classification type: text-classification dataset: name: emotion type: emotion config: split split: validation args: split metrics: - name: Accuracy type: accuracy value: 0.941 - name: F1 type: f1 value: 0.9411169346964399 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-uncased-with-preprocess-finetuned-emotion-5-epochs-5e-05-lr-0.1-weight_decay This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.2591 - Accuracy: 0.941 - F1: 0.9411 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.0799 | 1.0 | 250 | 0.1898 | 0.9375 | 0.9377 | | 0.0516 | 2.0 | 500 | 0.2290 | 0.938 | 0.9383 | | 0.0386 | 3.0 | 750 | 0.2107 | 0.9415 | 0.9419 | | 0.0195 | 4.0 | 1000 | 0.2607 | 0.9435 | 0.9433 | | 0.0149 | 5.0 | 1250 | 0.2591 | 0.941 | 0.9411 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3
TheBloke/robin-33B-v2-GGML
TheBloke
2023-06-16T23:31:16Z
0
5
null
[ "license:other", "region:us" ]
null
2023-06-16T18:09:39Z
--- inference: false license: other --- <!-- header start --> <div style="width: 100%;"> <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <div style="display: flex; justify-content: space-between; width: 100%;"> <div style="display: flex; flex-direction: column; align-items: flex-start;"> <p><a href="https://discord.gg/Jq4vkcDakD">Chat & support: my new Discord server</a></p> </div> <div style="display: flex; flex-direction: column; align-items: flex-end;"> <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p> </div> </div> <!-- header end --> # OptimalScale's Robin 33B v2 GGML These files are GGML format model files for [OptimalScale's Robin 33B v2](https://huggingface.co/OptimalScale/robin-33b-v2-delta). GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as: * [text-generation-webui](https://github.com/oobabooga/text-generation-webui) * [KoboldCpp](https://github.com/LostRuins/koboldcpp) * [ParisNeo/GPT4All-UI](https://github.com/ParisNeo/gpt4all-ui) * [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) * [ctransformers](https://github.com/marella/ctransformers) ## Repositories available * [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/robin-33B-v2-GPTQ) * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/robin-33B-v2-GGML) * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/robin-33B-v2-fp16) ## Prompt template ``` A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions ###Human: prompt ###Assistant: ``` <!-- compatibility_ggml start --> ## Compatibility ### Original llama.cpp quant methods: `q4_0, q4_1, q5_0, q5_1, q8_0` I have quantized these 'original' quantisation methods using an older version of llama.cpp so that they remain compatible with llama.cpp as of May 19th, commit `2d5db48`. These are guaranteed to be compatbile with any UIs, tools and libraries released since late May. ### New k-quant methods: `q2_K, q3_K_S, q3_K_M, q3_K_L, q4_K_S, q4_K_M, q5_K_S, q6_K` These new quantisation methods are compatible with llama.cpp as of June 6th, commit `2d43387`. They are now also compatible with recent releases of text-generation-webui, KoboldCpp, llama-cpp-python and ctransformers. Other tools and libraries may or may not be compatible - check their documentation if in doubt. ## Explanation of the new k-quant methods The new methods available are: * GGML_TYPE_Q2_K - "type-1" 2-bit quantization in super-blocks containing 16 blocks, each block having 16 weight. Block scales and mins are quantized with 4 bits. This ends up effectively using 2.5625 bits per weight (bpw) * GGML_TYPE_Q3_K - "type-0" 3-bit quantization in super-blocks containing 16 blocks, each block having 16 weights. Scales are quantized with 6 bits. This end up using 3.4375 bpw. * GGML_TYPE_Q4_K - "type-1" 4-bit quantization in super-blocks containing 8 blocks, each block having 32 weights. Scales and mins are quantized with 6 bits. This ends up using 4.5 bpw. * GGML_TYPE_Q5_K - "type-1" 5-bit quantization. Same super-block structure as GGML_TYPE_Q4_K resulting in 5.5 bpw * GGML_TYPE_Q6_K - "type-0" 6-bit quantization. Super-blocks with 16 blocks, each block having 16 weights. Scales are quantized with 8 bits. This ends up using 6.5625 bpw * GGML_TYPE_Q8_K - "type-0" 8-bit quantization. Only used for quantizing intermediate results. The difference to the existing Q8_0 is that the block size is 256. All 2-6 bit dot products are implemented for this quantization type. Refer to the Provided Files table below to see what files use which methods, and how. <!-- compatibility_ggml end --> ## Provided files | Name | Quant method | Bits | Size | Max RAM required | Use case | | ---- | ---- | ---- | ---- | ---- | ----- | | robin-33b.ggmlv3.q2_K.bin | q2_K | 2 | 13.71 GB | 16.21 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.vw and feed_forward.w2 tensors, GGML_TYPE_Q2_K for the other tensors. | | robin-33b.ggmlv3.q3_K_L.bin | q3_K_L | 3 | 17.28 GB | 19.78 GB | New k-quant method. Uses GGML_TYPE_Q5_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | robin-33b.ggmlv3.q3_K_M.bin | q3_K_M | 3 | 15.72 GB | 18.22 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | robin-33b.ggmlv3.q3_K_S.bin | q3_K_S | 3 | 14.06 GB | 16.56 GB | New k-quant method. Uses GGML_TYPE_Q3_K for all tensors | | robin-33b.ggmlv3.q4_0.bin | q4_0 | 4 | 18.30 GB | 20.80 GB | Original llama.cpp quant method, 4-bit. | | robin-33b.ggmlv3.q4_1.bin | q4_1 | 4 | 20.33 GB | 22.83 GB | Original llama.cpp quant method, 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models. | | robin-33b.ggmlv3.q4_K_M.bin | q4_K_M | 4 | 19.62 GB | 22.12 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q4_K | | robin-33b.ggmlv3.q4_K_S.bin | q4_K_S | 4 | 18.36 GB | 20.86 GB | New k-quant method. Uses GGML_TYPE_Q4_K for all tensors | | robin-33b.ggmlv3.q5_0.bin | q5_0 | 5 | 22.37 GB | 24.87 GB | Original llama.cpp quant method, 5-bit. Higher accuracy, higher resource usage and slower inference. | | robin-33b.ggmlv3.q5_1.bin | q5_1 | 5 | 24.40 GB | 26.90 GB | Original llama.cpp quant method, 5-bit. Even higher accuracy, resource usage and slower inference. | | robin-33b.ggmlv3.q5_K_M.bin | q5_K_M | 5 | 23.05 GB | 25.55 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q5_K | | robin-33b.ggmlv3.q5_K_S.bin | q5_K_S | 5 | 22.40 GB | 24.90 GB | New k-quant method. Uses GGML_TYPE_Q5_K for all tensors | | robin-33b.ggmlv3.q6_K.bin | q6_K | 6 | 26.69 GB | 29.19 GB | New k-quant method. Uses GGML_TYPE_Q8_K - 6-bit quantization - for all tensors | | robin-33b.ggmlv3.q8_0.bin | q8_0 | 8 | 34.56 GB | 37.06 GB | Original llama.cpp quant method, 8-bit. Almost indistinguishable from float16. High resource use and slow. Not recommended for most users. | **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead. ## How to run in `llama.cpp` I use the following command line; adjust for your tastes and needs: ``` ./main -t 10 -ngl 32 -m robin-33b.ggmlv3.q5_0.bin --color -c 2048 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.\n###Human: write a story about llamas\n###Assistant:" ``` Change `-t 10` to the number of physical CPU cores you have. For example if your system has 8 cores/16 threads, use `-t 8`. Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration. If you want to have a chat-style conversation, replace the `-p <PROMPT>` argument with `-i -ins` ## How to run in `text-generation-webui` Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md). <!-- footer start --> ## Discord For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/Jq4vkcDakD) ## Thanks, and how to contribute. Thanks to the [chirper.ai](https://chirper.ai) team! I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. * Patreon: https://patreon.com/TheBlokeAI * Ko-Fi: https://ko-fi.com/TheBlokeAI **Special thanks to**: Luke from CarbonQuill, Aemon Algiz, Dmitriy Samsonov. **Patreon special mentions**: vamX, K, Jonathan Leane, Lone Striker, Sean Connelly, Chris McCloskey, WelcomeToTheClub, Nikolai Manek, John Detwiler, Kalila, David Flickinger, Fen Risland, subjectnull, Johann-Peter Hartmann, Talal Aujan, John Villwock, senxiiz, Khalefa Al-Ahmad, Kevin Schuppel, Alps Aficionado, Derek Yates, Mano Prime, Nathan LeClaire, biorpg, trip7s trip, Asp the Wyvern, chris gileta, Iucharbius , Artur Olbinski, Ai Maven, Joseph William Delisle, Luke Pendergrass, Illia Dulskyi, Eugene Pentland, Ajan Kanaga, Willem Michiel, Space Cruiser, Pyrater, Preetika Verma, Junyu Yang, Oscar Rangel, Spiking Neurons AB, Pierre Kircher, webtim, Cory Kujawski, terasurfer , Trenton Dambrowitz, Gabriel Puliatti, Imad Khwaja, Luke. Thank you to all my generous patrons and donaters! <!-- footer end --> # Original model card: OptimalScale's Robin 33B v2 No model card provided in source repository.
TacLucas/Test
TacLucas
2023-06-16T23:07:19Z
0
0
null
[ "arxiv:1910.09700", "region:us" ]
null
2023-06-16T23:06:14Z
--- # For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1 # Doc / guide: https://huggingface.co/docs/hub/model-cards {} --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Data Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
ghze/Taxi_v3
ghze
2023-06-16T23:00:53Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T23:00:48Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: Taxi_v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.54 +/- 2.71 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="ghze/Taxi_v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
amjadfqs/finalProject
amjadfqs
2023-06-16T22:28:48Z
15
0
transformers
[ "transformers", "pytorch", "tensorboard", "swin", "image-classification", "generated_from_trainer", "dataset:imagefolder", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2023-06-15T17:30:57Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imagefolder metrics: - accuracy - precision model-index: - name: finalProject results: - task: name: Image Classification type: image-classification dataset: name: imagefolder type: imagefolder config: default split: train args: default metrics: - name: Accuracy type: accuracy value: 0.9890023566378633 - name: Precision type: precision value: 0.9894345375382527 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finalProject This model is a fine-tuned version of [microsoft/swin-base-patch4-window7-224-in22k](https://huggingface.co/microsoft/swin-base-patch4-window7-224-in22k) on the imagefolder dataset. It achieves the following results on the evaluation set: - Loss: 0.0411 - Accuracy: 0.9890 - F1 Score: 0.9892 - Precision: 0.9894 - Sensitivity: 0.9891 - Specificity: 0.9972 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 256 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 Score | Precision | Sensitivity | Specificity | |:-------------:|:-----:|:----:|:---------------:|:--------:|:--------:|:---------:|:-----------:|:-----------:| | 0.3384 | 1.0 | 30 | 0.2387 | 0.9144 | 0.9163 | 0.9197 | 0.9146 | 0.9781 | | 0.1608 | 2.0 | 60 | 0.1635 | 0.9466 | 0.9476 | 0.9485 | 0.9474 | 0.9865 | | 0.0953 | 3.0 | 90 | 0.0915 | 0.9698 | 0.9703 | 0.9706 | 0.9706 | 0.9924 | | 0.0573 | 4.0 | 120 | 0.1125 | 0.9607 | 0.9617 | 0.9634 | 0.9621 | 0.9901 | | 0.0335 | 5.0 | 150 | 0.0536 | 0.9827 | 0.9831 | 0.9837 | 0.9826 | 0.9957 | | 0.0185 | 6.0 | 180 | 0.0543 | 0.9827 | 0.9830 | 0.9837 | 0.9825 | 0.9957 | | 0.0226 | 7.0 | 210 | 0.0478 | 0.9859 | 0.9861 | 0.9866 | 0.9856 | 0.9965 | | 0.0131 | 8.0 | 240 | 0.0468 | 0.9843 | 0.9846 | 0.9847 | 0.9846 | 0.9961 | | 0.0087 | 9.0 | 270 | 0.0411 | 0.9890 | 0.9892 | 0.9894 | 0.9891 | 0.9972 | | 0.0043 | 10.0 | 300 | 0.0376 | 0.9886 | 0.9888 | 0.9890 | 0.9887 | 0.9971 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu117 - Datasets 2.13.0 - Tokenizers 0.13.3
devonho/my_awesome_opus_books_model
devonho
2023-06-16T22:28:30Z
8
0
transformers
[ "transformers", "pytorch", "t5", "text2text-generation", "generated_from_trainer", "dataset:opus100", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2023-06-06T07:28:09Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - opus100 metrics: - bleu model-index: - name: my_awesome_opus_books_model results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: opus100 type: opus100 config: en-ja split: test args: en-ja metrics: - name: Bleu type: bleu value: 23.8215 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_awesome_opus_books_model This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the opus100 dataset. It achieves the following results on the evaluation set: - Loss: 0.4506 - Bleu: 23.8215 - Gen Len: 4.6055 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len | |:-------------:|:-----:|:-------:|:---------------:|:-------:|:-------:| | 0.4468 | 1.0 | 500000 | 0.4585 | 23.9023 | 4.705 | | 0.4397 | 2.0 | 1000000 | 0.4506 | 23.8215 | 4.6055 | ### Framework versions - Transformers 4.29.2 - Pytorch 2.0.1+cu117 - Datasets 2.12.0 - Tokenizers 0.13.3
Enterprize1/ppo-LunarLander-v2
Enterprize1
2023-06-16T21:45:24Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T21:45:00Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 242.78 +/- 66.66 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
yinxiaoz/bert-finetuned-ner
yinxiaoz
2023-06-16T21:37:53Z
109
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "token-classification", "generated_from_trainer", "dataset:conll2003", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2023-06-15T05:15:04Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - conll2003 metrics: - precision - recall - f1 - accuracy model-index: - name: bert-finetuned-ner results: - task: name: Token Classification type: token-classification dataset: name: conll2003 type: conll2003 config: conll2003 split: validation args: conll2003 metrics: - name: Precision type: precision value: 0.9326065411298315 - name: Recall type: recall value: 0.9501851228542578 - name: F1 type: f1 value: 0.9413137712570858 - name: Accuracy type: accuracy value: 0.9867104256195914 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-finetuned-ner This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.0600 - Precision: 0.9326 - Recall: 0.9502 - F1: 0.9413 - Accuracy: 0.9867 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0884 | 1.0 | 1756 | 0.0675 | 0.9186 | 0.9339 | 0.9261 | 0.9822 | | 0.0345 | 2.0 | 3512 | 0.0611 | 0.9291 | 0.9485 | 0.9387 | 0.9862 | | 0.0182 | 3.0 | 5268 | 0.0600 | 0.9326 | 0.9502 | 0.9413 | 0.9867 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3
stanford-crfm/music-small-ar-inter-100k
stanford-crfm
2023-06-16T21:28:37Z
182
1
transformers
[ "transformers", "pytorch", "gpt2", "arxiv:2306.08620", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2023-06-05T00:04:27Z
--- license: apache-2.0 --- This is a Small (112M parameter) Transformer trained for 100k steps on interarrival-time encoded music from the [Lakh MIDI dataset](https://colinraffel.com/projects/lmd/). # References for the Anticipatory Music Transformer The Anticipatory Music Transformer paper is available on [ArXiv](http://arxiv.org/abs/2306.08620). The full model card is available [here](https://johnthickstun.com/assets/pdf/music-modelcard.pdf). Code for using this model is available on [GitHub](https://github.com/jthickstun/anticipation/). See the accompanying [blog post](https://crfm.stanford.edu/2023/06/16/anticipatory-music-transformer.html) for additional discussion of this model.
stanford-crfm/music-small-ar-100k
stanford-crfm
2023-06-16T21:27:39Z
184
0
transformers
[ "transformers", "pytorch", "gpt2", "arxiv:2306.08620", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2023-06-04T23:58:03Z
--- license: apache-2.0 --- This is a Small (128M parameter) Transformer trained for 100k steps on arrival-time encoded music from the [Lakh MIDI dataset](https://colinraffel.com/projects/lmd/). # References for the Anticipatory Music Transformer The Anticipatory Music Transformer paper is available on [ArXiv](http://arxiv.org/abs/2306.08620). The full model card is available [here](https://johnthickstun.com/assets/pdf/music-modelcard.pdf). Code for using this model is available on [GitHub](https://github.com/jthickstun/anticipation/). See the accompanying [blog post](https://crfm.stanford.edu/2023/06/16/anticipatory-music-transformer.html) for additional discussion of this model.
stanford-crfm/music-small-100k
stanford-crfm
2023-06-16T21:26:29Z
181
0
transformers
[ "transformers", "pytorch", "gpt2", "arxiv:2306.08620", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2023-06-04T23:26:52Z
--- license: apache-2.0 --- This is a Small (128M parameter) Transformer trained for 100k steps on arrival-time encoded music from the [Lakh MIDI dataset](https://colinraffel.com/projects/lmd/). This model was trained with anticipation. # References for the Anticipatory Music Transformer The Anticipatory Music Transformer paper is available on [ArXiv](http://arxiv.org/abs/2306.08620). The full model card is available [here](https://johnthickstun.com/assets/pdf/music-modelcard.pdf). Code for using this model is available on [GitHub](https://github.com/jthickstun/anticipation/). See the accompanying [blog post](https://crfm.stanford.edu/2023/06/16/anticipatory-music-transformer.html) for additional discussion of this model.
stanford-crfm/music-medium-800k
stanford-crfm
2023-06-16T21:25:52Z
572
4
transformers
[ "transformers", "pytorch", "gpt2", "arxiv:2306.08620", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2023-06-05T00:17:20Z
--- license: apache-2.0 --- This is a Medium (360M parameter) Transformer trained for 800k steps on arrival-time encoded music from the [Lakh MIDI dataset](https://colinraffel.com/projects/lmd/). This model was trained with anticipation. # References for the Anticipatory Music Transformer The Anticipatory Music Transformer paper is available on [ArXiv](http://arxiv.org/abs/2306.08620). The full model card is available [here](https://johnthickstun.com/assets/pdf/music-modelcard.pdf). Code for using this model is available on [GitHub](https://github.com/jthickstun/anticipation/). See the accompanying [blog post](https://crfm.stanford.edu/2023/06/16/anticipatory-music-transformer.html) for additional discussion of this model.
stanford-crfm/music-medium-200k
stanford-crfm
2023-06-16T21:25:22Z
180
0
transformers
[ "transformers", "pytorch", "gpt2", "arxiv:2306.08620", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2023-06-05T00:13:20Z
--- license: apache-2.0 --- This is a Medium (360M parameter) Transformer trained for 200k steps on arrival-time encoded music from the [Lakh MIDI dataset](https://colinraffel.com/projects/lmd/). This model was trained with anticipation. # References for the Anticipatory Music Transformer The Anticipatory Music Transformer paper is available on [ArXiv](http://arxiv.org/abs/2306.08620). The full model card is available [here](https://johnthickstun.com/assets/pdf/music-modelcard.pdf). Code for using this model is available on [GitHub](https://github.com/jthickstun/anticipation/). See the accompanying [blog post](https://crfm.stanford.edu/2023/06/16/anticipatory-music-transformer.html) for additional discussion of this model.
stanford-crfm/music-medium-100k
stanford-crfm
2023-06-16T21:24:54Z
176
0
transformers
[ "transformers", "pytorch", "gpt2", "arxiv:2306.08620", "license:apache-2.0", "text-generation-inference", "endpoints_compatible", "region:us" ]
null
2023-06-05T00:08:04Z
--- license: apache-2.0 --- This is a Medium (360M parameter) Transformer trained for 100k steps on arrival-time encoded music from the [Lakh MIDI dataset](https://colinraffel.com/projects/lmd/). This model was trained with anticipation. # References for the Anticipatory Music Transformer The Anticipatory Music Transformer paper is available on [ArXiv](http://arxiv.org/abs/2306.08620). The full model card is available [here](https://johnthickstun.com/assets/pdf/music-modelcard.pdf). Code for using this model is available on [GitHub](https://github.com/jthickstun/anticipation/). See the accompanying [blog post](https://crfm.stanford.edu/2023/06/16/anticipatory-music-transformer.html) for additional discussion of this model.
jondurbin/airoboros-65b-gpt4-1.2-peft
jondurbin
2023-06-16T21:01:26Z
0
0
null
[ "dataset:jondurbin/airoboros-gpt4-1.2", "license:other", "region:us" ]
null
2023-06-14T09:11:36Z
--- license: other datasets: - jondurbin/airoboros-gpt4-1.2 --- peft weights of https://hugginface.co/jondurbin/airoboros-65b-gpt4-1.2, see that card for details
crlandsc/bsrnn-bass
crlandsc
2023-06-16T20:24:33Z
0
1
null
[ "audio source separation", "music demixing", "band-split recurrent neural network", "bsrnn", "spectrogram", "bass", "region:us" ]
null
2023-06-16T20:16:53Z
--- tags: - audio source separation - music demixing - band-split recurrent neural network - bsrnn - spectrogram - bass --- # Model Card for bsrnn-bass Bass model for [Music-Demixing-with-Band-Split-RNN](https://github.com/crlandsc/Music-Demixing-with-Band-Split-RNN).
GEMCorp/q-Taxi-v3
GEMCorp
2023-06-16T20:19:33Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T20:08:42Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-Taxi-v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.56 +/- 2.71 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="GEMCorp/q-Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
sheshenin/shshnnphoto
sheshenin
2023-06-16T20:14:44Z
41
0
diffusers
[ "diffusers", "safetensors", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2023-06-16T20:09:15Z
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### Shshnnphoto Dreambooth model trained by sheshenin with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Sample pictures of this concept:
TheBloke/robin-13B-v2-GGML
TheBloke
2023-06-16T20:13:21Z
0
6
null
[ "license:other", "region:us" ]
null
2023-06-16T18:59:47Z
--- inference: false license: other --- <!-- header start --> <div style="width: 100%;"> <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <div style="display: flex; justify-content: space-between; width: 100%;"> <div style="display: flex; flex-direction: column; align-items: flex-start;"> <p><a href="https://discord.gg/Jq4vkcDakD">Chat & support: my new Discord server</a></p> </div> <div style="display: flex; flex-direction: column; align-items: flex-end;"> <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p> </div> </div> <!-- header end --> # OptimalScale's Robin 13B v2 GGML These files are GGML format model files for [OptimalScale's Robin 13B v2](https://huggingface.co/OptimalScale/robin-13b-v2-delta). GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as: * [text-generation-webui](https://github.com/oobabooga/text-generation-webui) * [KoboldCpp](https://github.com/LostRuins/koboldcpp) * [ParisNeo/GPT4All-UI](https://github.com/ParisNeo/gpt4all-ui) * [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) * [ctransformers](https://github.com/marella/ctransformers) ## Prompt template ``` A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions ###Human: prompt ###Assistant: ``` ## Repositories available * [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/robin-13B-v2-GPTQ) * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/robin-13B-v2-GGML) * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/robin-13B-v2-fp16) <!-- compatibility_ggml start --> ## Compatibility ### Original llama.cpp quant methods: `q4_0, q4_1, q5_0, q5_1, q8_0` I have quantized these 'original' quantisation methods using an older version of llama.cpp so that they remain compatible with llama.cpp as of May 19th, commit `2d5db48`. These are guaranteed to be compatbile with any UIs, tools and libraries released since late May. ### New k-quant methods: `q2_K, q3_K_S, q3_K_M, q3_K_L, q4_K_S, q4_K_M, q5_K_S, q6_K` These new quantisation methods are compatible with llama.cpp as of June 6th, commit `2d43387`. They are now also compatible with recent releases of text-generation-webui, KoboldCpp, llama-cpp-python and ctransformers. Other tools and libraries may or may not be compatible - check their documentation if in doubt. ## Explanation of the new k-quant methods The new methods available are: * GGML_TYPE_Q2_K - "type-1" 2-bit quantization in super-blocks containing 16 blocks, each block having 16 weight. Block scales and mins are quantized with 4 bits. This ends up effectively using 2.5625 bits per weight (bpw) * GGML_TYPE_Q3_K - "type-0" 3-bit quantization in super-blocks containing 16 blocks, each block having 16 weights. Scales are quantized with 6 bits. This end up using 3.4375 bpw. * GGML_TYPE_Q4_K - "type-1" 4-bit quantization in super-blocks containing 8 blocks, each block having 32 weights. Scales and mins are quantized with 6 bits. This ends up using 4.5 bpw. * GGML_TYPE_Q5_K - "type-1" 5-bit quantization. Same super-block structure as GGML_TYPE_Q4_K resulting in 5.5 bpw * GGML_TYPE_Q6_K - "type-0" 6-bit quantization. Super-blocks with 16 blocks, each block having 16 weights. Scales are quantized with 8 bits. This ends up using 6.5625 bpw * GGML_TYPE_Q8_K - "type-0" 8-bit quantization. Only used for quantizing intermediate results. The difference to the existing Q8_0 is that the block size is 256. All 2-6 bit dot products are implemented for this quantization type. Refer to the Provided Files table below to see what files use which methods, and how. <!-- compatibility_ggml end --> ## Provided files | Name | Quant method | Bits | Size | Max RAM required | Use case | | ---- | ---- | ---- | ---- | ---- | ----- | | robin-13b.ggmlv3.q2_K.bin | q2_K | 2 | 5.51 GB | 8.01 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.vw and feed_forward.w2 tensors, GGML_TYPE_Q2_K for the other tensors. | | robin-13b.ggmlv3.q3_K_L.bin | q3_K_L | 3 | 6.93 GB | 9.43 GB | New k-quant method. Uses GGML_TYPE_Q5_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | robin-13b.ggmlv3.q3_K_M.bin | q3_K_M | 3 | 6.31 GB | 8.81 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | robin-13b.ggmlv3.q3_K_S.bin | q3_K_S | 3 | 5.66 GB | 8.16 GB | New k-quant method. Uses GGML_TYPE_Q3_K for all tensors | | robin-13b.ggmlv3.q4_0.bin | q4_0 | 4 | 7.32 GB | 9.82 GB | Original llama.cpp quant method, 4-bit. | | robin-13b.ggmlv3.q4_1.bin | q4_1 | 4 | 8.14 GB | 10.64 GB | Original llama.cpp quant method, 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models. | | robin-13b.ggmlv3.q4_K_M.bin | q4_K_M | 4 | 7.87 GB | 10.37 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q4_K | | robin-13b.ggmlv3.q4_K_S.bin | q4_K_S | 4 | 7.37 GB | 9.87 GB | New k-quant method. Uses GGML_TYPE_Q4_K for all tensors | | robin-13b.ggmlv3.q5_0.bin | q5_0 | 5 | 8.95 GB | 11.45 GB | Original llama.cpp quant method, 5-bit. Higher accuracy, higher resource usage and slower inference. | | robin-13b.ggmlv3.q5_1.bin | q5_1 | 5 | 9.76 GB | 12.26 GB | Original llama.cpp quant method, 5-bit. Even higher accuracy, resource usage and slower inference. | | robin-13b.ggmlv3.q5_K_M.bin | q5_K_M | 5 | 9.23 GB | 11.73 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q5_K | | robin-13b.ggmlv3.q5_K_S.bin | q5_K_S | 5 | 8.97 GB | 11.47 GB | New k-quant method. Uses GGML_TYPE_Q5_K for all tensors | | robin-13b.ggmlv3.q6_K.bin | q6_K | 6 | 10.68 GB | 13.18 GB | New k-quant method. Uses GGML_TYPE_Q8_K - 6-bit quantization - for all tensors | | robin-13b.ggmlv3.q8_0.bin | q8_0 | 8 | 13.83 GB | 16.33 GB | Original llama.cpp quant method, 8-bit. Almost indistinguishable from float16. High resource use and slow. Not recommended for most users. | **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead. ## How to run in `llama.cpp` I use the following command line; adjust for your tastes and needs: ``` ./main -t 10 -ngl 32 -m robin-13b.ggmlv3.q5_0.bin --color -c 2048 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.\n###Human: write a story about llamas\n###Assistant:" ``` Change `-t 10` to the number of physical CPU cores you have. For example if your system has 8 cores/16 threads, use `-t 8`. Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration. If you want to have a chat-style conversation, replace the `-p <PROMPT>` argument with `-i -ins` ## How to run in `text-generation-webui` Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md). <!-- footer start --> ## Discord For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/Jq4vkcDakD) ## Thanks, and how to contribute. Thanks to the [chirper.ai](https://chirper.ai) team! I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. * Patreon: https://patreon.com/TheBlokeAI * Ko-Fi: https://ko-fi.com/TheBlokeAI **Special thanks to**: Luke from CarbonQuill, Aemon Algiz, Dmitriy Samsonov. **Patreon special mentions**: vamX, K, Jonathan Leane, Lone Striker, Sean Connelly, Chris McCloskey, WelcomeToTheClub, Nikolai Manek, John Detwiler, Kalila, David Flickinger, Fen Risland, subjectnull, Johann-Peter Hartmann, Talal Aujan, John Villwock, senxiiz, Khalefa Al-Ahmad, Kevin Schuppel, Alps Aficionado, Derek Yates, Mano Prime, Nathan LeClaire, biorpg, trip7s trip, Asp the Wyvern, chris gileta, Iucharbius , Artur Olbinski, Ai Maven, Joseph William Delisle, Luke Pendergrass, Illia Dulskyi, Eugene Pentland, Ajan Kanaga, Willem Michiel, Space Cruiser, Pyrater, Preetika Verma, Junyu Yang, Oscar Rangel, Spiking Neurons AB, Pierre Kircher, webtim, Cory Kujawski, terasurfer , Trenton Dambrowitz, Gabriel Puliatti, Imad Khwaja, Luke. Thank you to all my generous patrons and donaters! <!-- footer end --> # Original model card: OptimalScale's Robin 13B v2 No model card provided in source repository.
FALLENSTAR/MitsubishiChariotLoRa
FALLENSTAR
2023-06-16T20:10:47Z
0
0
null
[ "region:us" ]
null
2023-06-09T22:52:05Z
### Model Description That LoRa based on Mitsubishi Chariot/Chariot grandis 1997-2003. It's also a test model that poorly configured, so you have to play with the settings... The best images I was able to get with this LoRa were at these settings: Steps: 25 Sampler: DPM++ SDE Karras, CFG scale: 6.5 and with LoRa strength 0.8-1 ![00078-3304362297.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/2zwn_ybCAMQij0uC5C2xO.png) ![00080-176898055.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/3B-I5jhupVxoOCFgQVWln.png) ![00144-12952343.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/0Sqxm4quAS9P12NVxCVYG.png) ![00146-3385434376.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/jzItGhLhj3STdfdqHNgDv.png) ![00150-980315653.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/DYer7-yboQtaRymzhTZVe.png) ![00162-751177051.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/GYISC31YZ10YhLs8KHg8O.png) ![00096-2832495824.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/51Jc3IAV6rkNsUu-3NnQN.png) ![00002-3097255927.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/6m13zDtJSdK4syfujs2bu.png) ![00044-1201350781.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/11wdBPSt2rZljp2LrPFzV.png) ![00036-481276871.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/vf3A8yRQURMkXQw6beyBE.png) ![00035-307154204.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/Pfe5dCwI9t_qXZzuI08t3.png) ![00034-2484733566.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/4OmkzRvSUCRRO91Gz-nh4.png) ![00033-417478899.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/q8S6-lxV2uY3BwZ5WkcL5.png) ![00029-3305906991.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/DSZHwQN1jhA1leW2jcvG8.png) ![00024-2340357030.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/6-LTPowo0YXLtIpGpByc7.png) ![00018-4030886737.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/ybEN_cmAo6xkha5W8EVuF.png) ![00005-3932370153.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/JBxIq6dzqXgIyWPAawOQW.png)
FALLENSTAR/TurbofansLoRa
FALLENSTAR
2023-06-16T20:09:36Z
0
0
null
[ "region:us" ]
null
2023-06-11T01:32:23Z
### Model Description This LoRa is based on Turbofan or Aero Covers, an invention from Japan. Turbofan were created to effectively cool the brake discs. Originally they were used in motorsports, and were made out of aluminum. Now, thanks to new brake technology, Turbofans are not used for their original purpose. And they are not popular in professional motorsports. But, to me, they add a futuristic style to car tuning. The best images I was able to get with this LoRa were at these settings: Steps: 25 Sampler: DPM++ SDE Karras, CFG scale: 6.5 and with LoRa strength 0.8-1 ![00094-3727343714.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/t4pEcqkV-1UZzNmVS8G5s.png) ![00096-3846199597.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/D0Wh8K3xII6uhsdaMpm_T.png) ![00112-2164319160.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/GUKuqj2iMjHGrh1ByuNkG.png) ![00117-4155383315.png](https://s3.amazonaws.com/moonup/production/uploads/640761fd2e309e65452364fa/1E5xBqoBmnB8YZfr32mtU.png)
apopam/Taxi-v3
apopam
2023-06-16T19:55:18Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T19:55:12Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: Taxi-v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.52 +/- 2.73 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="apopam/Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
GEMCorp/q-FrozenLake-v1-4x4-noSlippery
GEMCorp
2023-06-16T19:51:18Z
0
0
null
[ "FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T19:51:12Z
--- tags: - FrozenLake-v1-4x4-no_slippery - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-FrozenLake-v1-4x4-noSlippery results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: FrozenLake-v1-4x4-no_slippery type: FrozenLake-v1-4x4-no_slippery metrics: - type: mean_reward value: 1.00 +/- 0.00 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="GEMCorp/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
ananay/kneearch
ananay
2023-06-16T19:17:59Z
22
0
diffusers
[ "diffusers", "safetensors", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2023-06-16T19:05:11Z
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### kneearch Dreambooth model trained by ananay with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Sample pictures of this concept:
apopam/q-FrozenLake-v1-4x4-noSlippery
apopam
2023-06-16T18:55:01Z
0
0
null
[ "FrozenLake-v1-4x4", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T18:54:53Z
--- tags: - FrozenLake-v1-4x4 - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-FrozenLake-v1-4x4-noSlippery results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: FrozenLake-v1-4x4 type: FrozenLake-v1-4x4 metrics: - type: mean_reward value: 0.64 +/- 0.48 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="apopam/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
digiplay/majicMIX_realistic_v5preview
digiplay
2023-06-16T18:49:48Z
397
1
diffusers
[ "diffusers", "safetensors", "license:other", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2023-06-14T13:09:24Z
--- license: other --- Very famous realistic beauty Model Model info : https://civitai.com/models/43331?modelVersionId=79068 Orginal Author's DEMO image : ![](https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/09b88ff6-1e5b-4b15-82df-7e458d127b55/width=1024/00047-3896603326.jpeg)
Bodolaz/unit-1
Bodolaz
2023-06-16T18:46:03Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T18:45:38Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 249.33 +/- 43.84 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
Trisert/RedPajama-INCITE-Base-3B-v1-wikipedia
Trisert
2023-06-16T18:45:54Z
0
0
null
[ "generated_from_trainer", "license:apache-2.0", "region:us" ]
null
2023-06-16T18:40:38Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: RedPajama-INCITE-Base-3B-v1-wikipedia results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # RedPajama-INCITE-Base-3B-v1-wikipedia This model is a fine-tuned version of [togethercomputer/RedPajama-INCITE-Base-3B-v1](https://huggingface.co/togethercomputer/RedPajama-INCITE-Base-3B-v1) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 1 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2 - training_steps: 10 ### Training results ### Framework versions - Transformers 4.31.0.dev0 - Pytorch 2.0.0 - Datasets 2.13.0 - Tokenizers 0.13.3
JvThunder/ppo-Pyramids
JvThunder
2023-06-16T18:37:51Z
1
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
2023-06-16T18:37:41Z
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: JvThunder/ppo-Pyramids 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
davidmunechika/coreml-dreamlike-diffusion-1.0-palletized
davidmunechika
2023-06-16T18:36:11Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T18:33:59Z
--- license: creativeml-openrail-m ---
Fred01/ppo-LunarLander-v2
Fred01
2023-06-16T18:27:12Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T18:26:43Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 251.34 +/- 29.23 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
sambanovasystems/codegen-16B-mono-toolbench
sambanovasystems
2023-06-16T18:24:41Z
16
5
transformers
[ "transformers", "pytorch", "codegen", "text-generation", "arxiv:2305.16504", "arxiv:2203.13474", "license:bsd-3-clause", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2023-05-22T16:48:51Z
--- license: bsd-3-clause --- # codgen-16B-mono-toolbench <!-- Provide a quick summary of what the model is/does. --> codgen-16B-mono-toolbench is a 16 billion parameter model used for api based action generation. It is instruction tuned from [codegen-16B-mono](https://huggingface.co/Salesforce/codegen-16B-mono) on api based action generation datasets. ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [SambaNova Systems](https://sambanova.ai/) - **Model type:** Language Model - **Language(s):** English - **License:** bsd-3-clause - **Finetuned from model:** [codegen-16B-mono](https://huggingface.co/Salesforce/codegen-16B-mono) ### Basic Information <!-- Provide the basic links for the model. --> - **Paper**: [Link](https://arxiv.org/abs/2305.16504) - **Github**: [link](https://github.com/sambanova/toolbench) ## Uses <details> <summary>Click to expand</summary> <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> This model is intended for commercial and research use. ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> codgen-16B-mono-toolbench should NOT be used for purpose other than API based action generation. </details> --- ## How to Get Started with the Model <details> <summary>Click to expand</summary> ### Loading in model with Huggingface ```python from transformers import AutoModelForCausalLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("sambanovasystems/codegen-16B-mono-toolbench") model = AutoModelForCausalLM.from_pretrained("sambanovasystems/codegen-16B-mono-toolbench", device_map="auto", torch_dtype="auto") ``` ### Suggested Inference Parameters - do_sample: False ### Example Prompts To Try in GPU Tutorial Prompt 1: ``` I have the following set of API:\n\n# To set the maximum commute time in minute to your office location, assuming the office location is already defined\nAPI.set_max_commute_time(value: int)\n\n# To set the maximum home size in square feet\nAPI.set_max_square_feet(value: int)\n\n# To set the minimum home price in dollars\nAPI.set_min_price(value: int)\n\n# To set the number of garage(s)\nAPI.set_num_garages(value: int)\n\n# To set home types for search. For home buying, home_types choices are: \"House\", \"Townhouse\", \"Condo\", \"Land\", \"Multi-family\", \"Mobile\", \"Co-op\"; for home renting, home_types choices are: \"House\", \"Townhouse\", \"Condo\", \"Apartment\".\nAPI.select_home_type(home_types: List[str])\n\n# To set the number of balconies\nAPI.set_num_balconies(value: int)\n\n# Submit criterion to get search results. This function should be called after setting all the criterion.\nAPI.search()\n\n# To set the floor number\nAPI.set_floor_number(value: int)\n\n# To set the number of bedroom(s)\nAPI.set_num_beds(value: int)\n\n# To set the number of swimming pool(s)\nAPI.set_num_swimming_pools(value: int)\n\n# To set the maximum home price in dollars\nAPI.set_max_price(value: int)\n\n# To specify whether to search homes for buying or renting. 'value' can be chosen from ['buy', 'rent']. This function must be called after setting the location and before setting any other criteria.\nAPI.set_buy_or_rent(value: str)\n\n# To set the number of bathroom(s)\nAPI.set_num_baths(value: float)\n\n# To set the location for the search area. This function must be called before setting any criteria.\nAPI.set_location(value: string)\n\n# To set the minimum home size in square feet\nAPI.set_min_square_feet(value: int)\n\n-------------\n\nTask: Looking for homes to rent in Santa Clarita with a price range between $110000 and $1753000, a minimum of 1700 square feet, at least 2 balconies, and 3.5 bathrooms.\nAction:\n ``` Prompt 2: ``` I have the following set of API:\n\n# To set the location for hotel search, given a Loc object. This function must be called if booking type is 'hotels' or 'both'.\nAPI.set_hotel_location(Loc)\n\n# To set the number of hotel rooms to book.\nAPI.set_num_rooms(value)\n\n# To set the location for departure, given a Loc object. This function must be called if booking type is 'trip tickets' or 'both'.\nAPI.set_origin(Loc)\n\n# To select the transportation type from ['flight', 'train', 'bus', 'cruise']. This function must be called if booking type is 'trip tickets' or 'both'.\nAPI.select_transportation(transportation_type)\n\n# To set the return date of the trip, given a Date object. If booking type is 'both' and this function is not called explicitly, 'return_date' will be set to 'hotel_checkout_date' implicitly.\nAPI.set_return_date(Date)\n\n# To set the hotel check-in date, given a Date object. This function must be called if booking type is 'hotels' or 'both'.\nAPI.set_checkin_date(Date)\n\n# To define a date.\ndate = Date(month, day, year)\n\n# To set the departure date of the trip, given a Date object. This function must be called if booking type is 'trip tickets'. If booking type is 'both' and this function is not called explicitly, 'departure_date' will be set to 'hotel_checkin_date' implicitly.\nAPI.set_departure_date(Date)\n\n# To set the location for arrival, given a Loc object. This function must be called if booking type is 'trip tickets' or 'both'.\nAPI.set_destination(Loc)\n\n# To define a location of a given city 'City'.\nlocation = Loc('City')\n\n# To set maximum hotel room price.\nAPI.set_max_room_price(value)\n\n# To set minimum ticket price.\nAPI.set_min_ticket_price(value)\n\n# To select the booking type from ['hotels', 'trip tickets', 'both']. This function must be called before setting any criteria.\nAPI.select_booking_type(booking_type)\n\n# To set minimum hotel room price.\nAPI.set_min_room_price(value)\n\n# To set the number of child tickets to purchase.\nAPI.set_num_children(value)\n\n# To set the number of adult tickets to purchase.\nAPI.set_num_adults(value)\n\n# To select the hotel room type from ['King Bed', 'Queen Bed', 'Double', 'Luxury'].\nAPI.select_room_type(room_type)\n\n# To set maximum ticket price.\nAPI.set_max_ticket_price(value)\n\n# Submit criterion to get search results. This function should be called after setting all the criterion.\nAPI.search()\n\n# To set the hotel check-out date, given a Date object. This function must be called if booking type is 'hotels' or 'both'.\nAPI.set_checkout_date(Date)\n\n-------------\n\nTask: Looking to book 2 adult and 4 child tickets from Stockton to Baltimore by cruise, on 2023-07-29.\nAction:\n ``` </details> --- ## Training Details <details> <summary>Click to expand</summary> ### Training Data <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> The training data is curated for the 8 tasks in ToolBench. See Appendix A of the [paper](https://arxiv.org/abs/2305.16504) for task details and Appendix C.1 for the training data curation details. In total, there are 9704 training samples, organized in all-shot format as described in Appendix C.2. Here is the [download link](https://drive.google.com/file/d/1lUatLGnSVhfy1uVIPEQ7qCoLtnCIXi2O/view?usp=sharing) to the training data. ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> We trained codegen-16b-mono-toolbench on 4 80GB A100 gpu's. We started from [codegen-16B-mono](https://huggingface.co/Salesforce/codegen-16B-mono) and finetuned it on the dataset mentioned above. ### Hyperparameters - Hardware: A100 GPU - Optimizer: AdamW - Grad accumulation: 1 - Epochs: 8 - Global Batch size: 16 - Batch tokens: 16 * 2048 = 32,768 tokens - Learning Rate: 1e-5 - Learning Rate Scheduler: Fixed LR - Weight decay: 0.1 </details> ## Acknowledgment We would like to express our gratitude to the great work done in [CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis](https://arxiv.org/abs/2203.13474) ## Cite codegen-16b-mono-toolbench ``` @misc{xu2023tool, title={On the Tool Manipulation Capability of Open-source Large Language Models}, author={Qiantong Xu and Fenglu Hong and Bo Li and Changran Hu and Zhengyu Chen and Jian Zhang}, year={2023}, eprint={2305.16504}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
sambanovasystems/starcoder-toolbench
sambanovasystems
2023-06-16T18:23:22Z
23
4
transformers
[ "transformers", "pytorch", "gpt_bigcode", "text-generation", "arxiv:2305.16504", "arxiv:2305.06161", "license:bigcode-openrail-m", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2023-05-24T04:41:14Z
--- license: bigcode-openrail-m --- # starcoder-toolbench <!-- Provide a quick summary of what the model is/does. --> starcoder-toolbench is a 15 billion parameter model used for api based action generation. It is instruction tuned from [starcoder](https://huggingface.co/bigcode/starcoder) on api based action generation datasets. ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [SambaNova Systems](https://sambanova.ai/) - **Model type:** Language Model - **Language(s):** English - **License:** bigcode-openrail-m - **Finetuned from model:** [starcoder](https://huggingface.co/bigcode/starcoder) ### Basic Information <!-- Provide the basic links for the model. --> - **Paper**: [link](https://arxiv.org/abs/2305.16504) - **Github**: [link](https://github.com/sambanova/toolbench) ## Uses <details> <summary>Click to expand</summary> <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> This model is intended for commercial and research use. ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> starcoder-toolbench should NOT be used for purpose other than API based action generation. </details> --- ## How to Get Started with the Model <details> <summary>Click to expand</summary> ### Loading in model with Huggingface ```python from transformers import AutoModelForCausalLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("sambanovasystems/starcoder-toolbench") model = AutoModelForCausalLM.from_pretrained("sambanovasystems/starcoder-toolbench", device_map="auto", torch_dtype="auto") ``` ### Example Prompts To Try in GPU Tutorial Prompt 1: ``` I have the following set of API:\n\n# To set the maximum commute time in minute to your office location, assuming the office location is already defined\nAPI.set_max_commute_time(value: int)\n\n# To set the maximum home size in square feet\nAPI.set_max_square_feet(value: int)\n\n# To set the minimum home price in dollars\nAPI.set_min_price(value: int)\n\n# To set the number of garage(s)\nAPI.set_num_garages(value: int)\n\n# To set home types for search. For home buying, home_types choices are: \"House\", \"Townhouse\", \"Condo\", \"Land\", \"Multi-family\", \"Mobile\", \"Co-op\"; for home renting, home_types choices are: \"House\", \"Townhouse\", \"Condo\", \"Apartment\".\nAPI.select_home_type(home_types: List[str])\n\n# To set the number of balconies\nAPI.set_num_balconies(value: int)\n\n# Submit criterion to get search results. This function should be called after setting all the criterion.\nAPI.search()\n\n# To set the floor number\nAPI.set_floor_number(value: int)\n\n# To set the number of bedroom(s)\nAPI.set_num_beds(value: int)\n\n# To set the number of swimming pool(s)\nAPI.set_num_swimming_pools(value: int)\n\n# To set the maximum home price in dollars\nAPI.set_max_price(value: int)\n\n# To specify whether to search homes for buying or renting. 'value' can be chosen from ['buy', 'rent']. This function must be called after setting the location and before setting any other criteria.\nAPI.set_buy_or_rent(value: str)\n\n# To set the number of bathroom(s)\nAPI.set_num_baths(value: float)\n\n# To set the location for the search area. This function must be called before setting any criteria.\nAPI.set_location(value: string)\n\n# To set the minimum home size in square feet\nAPI.set_min_square_feet(value: int)\n\n-------------\n\nTask: Looking for homes to rent in Santa Clarita with a price range between $110000 and $1753000, a minimum of 1700 square feet, at least 2 balconies, and 3.5 bathrooms.\nAction:\n ``` Prompt 2: ``` I have the following set of API:\n\n# To set the location for hotel search, given a Loc object. This function must be called if booking type is 'hotels' or 'both'.\nAPI.set_hotel_location(Loc)\n\n# To set the number of hotel rooms to book.\nAPI.set_num_rooms(value)\n\n# To set the location for departure, given a Loc object. This function must be called if booking type is 'trip tickets' or 'both'.\nAPI.set_origin(Loc)\n\n# To select the transportation type from ['flight', 'train', 'bus', 'cruise']. This function must be called if booking type is 'trip tickets' or 'both'.\nAPI.select_transportation(transportation_type)\n\n# To set the return date of the trip, given a Date object. If booking type is 'both' and this function is not called explicitly, 'return_date' will be set to 'hotel_checkout_date' implicitly.\nAPI.set_return_date(Date)\n\n# To set the hotel check-in date, given a Date object. This function must be called if booking type is 'hotels' or 'both'.\nAPI.set_checkin_date(Date)\n\n# To define a date.\ndate = Date(month, day, year)\n\n# To set the departure date of the trip, given a Date object. This function must be called if booking type is 'trip tickets'. If booking type is 'both' and this function is not called explicitly, 'departure_date' will be set to 'hotel_checkin_date' implicitly.\nAPI.set_departure_date(Date)\n\n# To set the location for arrival, given a Loc object. This function must be called if booking type is 'trip tickets' or 'both'.\nAPI.set_destination(Loc)\n\n# To define a location of a given city 'City'.\nlocation = Loc('City')\n\n# To set maximum hotel room price.\nAPI.set_max_room_price(value)\n\n# To set minimum ticket price.\nAPI.set_min_ticket_price(value)\n\n# To select the booking type from ['hotels', 'trip tickets', 'both']. This function must be called before setting any criteria.\nAPI.select_booking_type(booking_type)\n\n# To set minimum hotel room price.\nAPI.set_min_room_price(value)\n\n# To set the number of child tickets to purchase.\nAPI.set_num_children(value)\n\n# To set the number of adult tickets to purchase.\nAPI.set_num_adults(value)\n\n# To select the hotel room type from ['King Bed', 'Queen Bed', 'Double', 'Luxury'].\nAPI.select_room_type(room_type)\n\n# To set maximum ticket price.\nAPI.set_max_ticket_price(value)\n\n# Submit criterion to get search results. This function should be called after setting all the criterion.\nAPI.search()\n\n# To set the hotel check-out date, given a Date object. This function must be called if booking type is 'hotels' or 'both'.\nAPI.set_checkout_date(Date)\n\n-------------\n\nTask: Looking to book 2 adult and 4 child tickets from Stockton to Baltimore by cruise, on 2023-07-29.\nAction:\n ``` </details> --- ## Training Details <details> <summary>Click to expand</summary> ### Training Data <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> The training data is curated for the 8 tasks in ToolBench. See Appendix A of the [paper](https://arxiv.org/abs/2305.16504) for task details and Appendix C.1 for the training data curation details. In total, there are 9704 training samples, organized in all-shot format as described in Appendix C.2. Here is the [download link](https://drive.google.com/file/d/1lUatLGnSVhfy1uVIPEQ7qCoLtnCIXi2O/view?usp=sharing) to the training data. ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> We trained starcoder-toolbench on 4 80GB A100 gpu's. We started from [starcoder](https://huggingface.co/bigcode/starcoder) and finetuned it on the dataset mentioned above. ### Hyperparameters - Hardware: A100 GPU - Optimizer: AdamW - Grad accumulation: 1 - Epochs: 8 - Global Batch size: 16 - Batch tokens: 16 * 2048 = 32,768 tokens - Learning Rate: 1e-5 - Learning Rate Scheduler: Fixed LR - Weight decay: 0.1 </details> ## Acknowledgment We would like to express our gratitude to the great work done in [StarCoder: may the source be with you!](https://arxiv.org/abs/2305.06161) ## Cite starcoder-toolbench ``` @misc{xu2023tool, title={On the Tool Manipulation Capability of Open-source Large Language Models}, author={Qiantong Xu and Fenglu Hong and Bo Li and Changran Hu and Zhengyu Chen and Jian Zhang}, year={2023}, eprint={2305.16504}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
ugiugi/inisw08-RoBERT-mlm-adamw_torch_bs8
ugiugi
2023-06-16T18:01:51Z
24
0
transformers
[ "transformers", "pytorch", "tensorboard", "roberta", "fill-mask", "generated_from_trainer", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2023-06-16T03:23:35Z
--- license: mit tags: - generated_from_trainer metrics: - accuracy model-index: - name: inisw08-RoBERT-mlm-adamw_torch_bs8 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # inisw08-RoBERT-mlm-adamw_torch_bs8 This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 4.4931 - Accuracy: 0.3551 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results ### Framework versions - Transformers 4.31.0.dev0 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3
gokuls/add_BERT_24_mnli
gokuls
2023-06-16T17:59:45Z
35
0
transformers
[ "transformers", "pytorch", "tensorboard", "hybridbert", "text-classification", "generated_from_trainer", "en", "dataset:glue", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-06-16T12:10:28Z
--- language: - en tags: - generated_from_trainer datasets: - glue metrics: - accuracy model-index: - name: add_BERT_24_mnli results: - task: name: Text Classification type: text-classification dataset: name: GLUE MNLI type: glue config: mnli split: validation_matched args: mnli metrics: - name: Accuracy type: accuracy value: 0.3295362082994304 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # add_BERT_24_mnli This model is a fine-tuned version of [gokuls/add_bert_12_layer_model_complete_training_new](https://huggingface.co/gokuls/add_bert_12_layer_model_complete_training_new) on the GLUE MNLI dataset. It achieves the following results on the evaluation set: - Loss: 1.0986 - Accuracy: 0.3295 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 4e-05 - train_batch_size: 128 - eval_batch_size: 128 - seed: 10 - distributed_type: multi-GPU - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 1.1032 | 1.0 | 3068 | 1.0994 | 0.3182 | | 1.0988 | 2.0 | 6136 | 1.0986 | 0.3182 | | 1.0987 | 3.0 | 9204 | 1.0987 | 0.3274 | | 1.0988 | 4.0 | 12272 | 1.0986 | 0.3274 | | 1.0987 | 5.0 | 15340 | 1.0986 | 0.3274 | | 1.0986 | 6.0 | 18408 | 1.0986 | 0.3182 | | 1.0986 | 7.0 | 21476 | 1.0986 | 0.3182 | | 1.0986 | 8.0 | 24544 | 1.0986 | 0.3182 | | 1.0986 | 9.0 | 27612 | 1.0986 | 0.3274 | ### Framework versions - Transformers 4.30.2 - Pytorch 1.14.0a0+410ce96 - Datasets 2.13.0 - Tokenizers 0.13.3
Trisert/outputs
Trisert
2023-06-16T17:43:33Z
0
0
null
[ "generated_from_trainer", "license:apache-2.0", "region:us" ]
null
2023-06-16T17:42:50Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: outputs results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # outputs This model is a fine-tuned version of [togethercomputer/RedPajama-INCITE-Base-3B-v1](https://huggingface.co/togethercomputer/RedPajama-INCITE-Base-3B-v1) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 1 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2 - training_steps: 10 ### Training results ### Framework versions - Transformers 4.31.0.dev0 - Pytorch 2.0.0 - Datasets 2.13.0 - Tokenizers 0.13.3
anirban-1009/tomato
anirban-1009
2023-06-16T17:34:28Z
0
0
keras
[ "keras", "en", "dataset:rotten_tomatoes", "license:agpl-3.0", "region:us" ]
null
2023-06-16T17:33:01Z
--- license: agpl-3.0 datasets: - rotten_tomatoes language: - en metrics: - accuracy library_name: keras ---
Ahatsham/flan-t5-small-imdb-text-classification
Ahatsham
2023-06-16T17:29:33Z
101
0
transformers
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2023-06-16T14:54:03Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: flan-t5-small-imdb-text-classification results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # flan-t5-small-imdb-text-classification This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results ### Framework versions - Transformers 4.28.1 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3
LarryAIDraw/chara_GrandBlue_KotegawaNanaka_v1
LarryAIDraw
2023-06-16T17:24:31Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T17:17:05Z
--- license: creativeml-openrail-m --- https://civitai.com/models/90866/kotegawa-nanaka-or-grand-blue
LarryAIDraw/chara_GrandBlue_KotegawaChisa_v1
LarryAIDraw
2023-06-16T17:24:20Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T17:16:42Z
--- license: creativeml-openrail-m --- https://civitai.com/models/90910/kotegawa-chisa-or-grand-blue
LarryAIDraw/jingliu100
LarryAIDraw
2023-06-16T17:23:29Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T17:14:17Z
--- license: creativeml-openrail-m --- https://civitai.com/models/89766/jingliu-or-honkai-star-rail
LarryAIDraw/mitsuki_nase-07
LarryAIDraw
2023-06-16T17:23:09Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T17:13:47Z
--- license: creativeml-openrail-m --- https://civitai.com/models/90270/mitsuki-nase-kyoukai-no-kanata
LarryAIDraw/shizuru-000010
LarryAIDraw
2023-06-16T17:21:46Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T17:12:22Z
--- license: creativeml-openrail-m --- https://civitai.com/models/90583/hoshino-shizurusummerswimsuit-or-princess-connect
LarryAIDraw/ganbaredouki-chan_douki-chan-11
LarryAIDraw
2023-06-16T17:20:27Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T17:11:48Z
--- license: creativeml-openrail-m --- https://civitai.com/models/89745/douki-chan-or-do-your-best-doki-chan
sd-dreambooth-library/BaysaLaban123
sd-dreambooth-library
2023-06-16T17:15:24Z
24
0
diffusers
[ "diffusers", "tensorboard", "text-to-image", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2023-06-16T17:13:19Z
--- license: creativeml-openrail-m tags: - text-to-image widget: - text: LabanBaysa1 --- ### Labanbaysa11 Dreambooth model trained by LabanAsmar with [Hugging Face Dreambooth Training Space](https://huggingface.co/spaces/multimodalart/dreambooth-training) with the v2-1-768 base model You run your new concept via `diffusers` [Colab Notebook for Inference](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_dreambooth_inference.ipynb). Don't forget to use the concept prompts! Sample pictures of: LabanBaysa1 (use that on your prompt) ![LabanBaysa1 0](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%281%29.jpg)![LabanBaysa1 1](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%282%29.jpg)![LabanBaysa1 2](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%283%29.jpg)![LabanBaysa1 3](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%284%29.jpg)![LabanBaysa1 4](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%285%29.jpg)![LabanBaysa1 5](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%286%29.jpg)![LabanBaysa1 6](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%287%29.jpg)![LabanBaysa1 7](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%288%29.jpg)![LabanBaysa1 8](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%289%29.jpg)![LabanBaysa1 9](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%2810%29.jpg)![LabanBaysa1 10](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%2811%29.jpg)![LabanBaysa1 11](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%2812%29.jpg)![LabanBaysa1 12](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%2813%29.jpg)![LabanBaysa1 13](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%2814%29.jpg)![LabanBaysa1 14](https://huggingface.co/LabanAsmar/labanbaysa11/resolve/main/concept_images/LabanBaysa1_%2815%29.jpg)
jncraton/flan-alpaca-xl-ct2-int8
jncraton
2023-06-16T16:29:38Z
12
0
transformers
[ "transformers", "dataset:tatsu-lab/alpaca", "arxiv:2306.04757", "arxiv:2210.11416", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2023-06-16T16:08:36Z
--- license: apache-2.0 datasets: - tatsu-lab/alpaca --- ## 🍮 🦙 Flan-Alpaca: Instruction Tuning from Humans and Machines 📣 Curious to know the performance of 🍮 🦙 **Flan-Alpaca** on large-scale LLM evaluation benchmark, **InstructEval**? Read our paper [https://arxiv.org/pdf/2306.04757.pdf](https://arxiv.org/pdf/2306.04757.pdf). We evaluated more than 10 open-source instruction-tuned LLMs belonging to various LLM families including Pythia, LLaMA, T5, UL2, OPT, and Mosaic. Codes and datasets: [https://github.com/declare-lab/instruct-eval](https://github.com/declare-lab/instruct-eval) 📣 **FLAN-T5** is also useful in text-to-audio generation. Find our work at [https://github.com/declare-lab/tango](https://github.com/declare-lab/tango) if you are interested. Our [repository](https://github.com/declare-lab/flan-alpaca) contains code for extending the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) synthetic instruction tuning to existing instruction-tuned models such as [Flan-T5](https://arxiv.org/abs/2210.11416). We have a [live interactive demo](https://huggingface.co/spaces/joaogante/transformers_streaming) thanks to [Joao Gante](https://huggingface.co/joaogante)! We are also benchmarking many instruction-tuned models at [declare-lab/flan-eval](https://github.com/declare-lab/flan-eval). Our pretrained models are fully available on HuggingFace 🤗 : | Model | Parameters | Instruction Data | Training GPUs | |----------------------------------------------------------------------------------|------------|----------------------------------------------------------------------------------------------------------------------------------------------------|-----------------| | [Flan-Alpaca-Base](https://huggingface.co/declare-lab/flan-alpaca-base) | 220M | [Flan](https://github.com/google-research/FLAN), [Alpaca](https://github.com/tatsu-lab/stanford_alpaca) | 1x A6000 | | [Flan-Alpaca-Large](https://huggingface.co/declare-lab/flan-alpaca-large) | 770M | [Flan](https://github.com/google-research/FLAN), [Alpaca](https://github.com/tatsu-lab/stanford_alpaca) | 1x A6000 | | [Flan-Alpaca-XL](https://huggingface.co/declare-lab/flan-alpaca-xl) | 3B | [Flan](https://github.com/google-research/FLAN), [Alpaca](https://github.com/tatsu-lab/stanford_alpaca) | 1x A6000 | | [Flan-Alpaca-XXL](https://huggingface.co/declare-lab/flan-alpaca-xxl) | 11B | [Flan](https://github.com/google-research/FLAN), [Alpaca](https://github.com/tatsu-lab/stanford_alpaca) | 4x A6000 (FSDP) | | [Flan-GPT4All-XL](https://huggingface.co/declare-lab/flan-gpt4all-xl) | 3B | [Flan](https://github.com/google-research/FLAN), [GPT4All](https://github.com/nomic-ai/gpt4all) | 1x A6000 | | [Flan-ShareGPT-XL](https://huggingface.co/declare-lab/flan-sharegpt-xl) | 3B | [Flan](https://github.com/google-research/FLAN), [ShareGPT](https://github.com/domeccleston/sharegpt)/[Vicuna](https://github.com/lm-sys/FastChat) | 1x A6000 | | [Flan-Alpaca-GPT4-XL*](https://huggingface.co/declare-lab/flan-alpaca-gpt4-xl) | 3B | [Flan](https://github.com/google-research/FLAN), [GPT4-Alpaca](https://github.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM) | 1x A6000 | *recommended for better performance ### Why? [Alpaca](https://crfm.stanford.edu/2023/03/13/alpaca.html) represents an exciting new direction to approximate the performance of large language models (LLMs) like ChatGPT cheaply and easily. Concretely, they leverage an LLM such as GPT-3 to generate instructions as synthetic training data. The synthetic data which covers more than 50k tasks can then be used to finetune a smaller model. However, the original implementation is less accessible due to licensing constraints of the underlying [LLaMA](https://ai.facebook.com/blog/large-language-model-llama-meta-ai/) model. Furthermore, users have noted [potential noise](https://github.com/tloen/alpaca-lora/issues/65) in the synthetic dataset. Hence, it may be better to explore a fully accessible model that is already trained on high-quality (but less diverse) instructions such as [Flan-T5](https://arxiv.org/abs/2210.11416). ### Usage ``` from transformers import pipeline prompt = "Write an email about an alpaca that likes flan" model = pipeline(model="declare-lab/flan-alpaca-gpt4-xl") model(prompt, max_length=128, do_sample=True) # Dear AlpacaFriend, # My name is Alpaca and I'm 10 years old. # I'm excited to announce that I'm a big fan of flan! # We like to eat it as a snack and I believe that it can help with our overall growth. # I'd love to hear your feedback on this idea. # Have a great day! # Best, AL Paca ```
YvesP/a2c-AntBulletEnv-v0
YvesP
2023-06-16T16:25:24Z
0
0
stable-baselines3
[ "stable-baselines3", "AntBulletEnv-v0", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T16:23:58Z
--- library_name: stable-baselines3 tags: - AntBulletEnv-v0 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: A2C results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: AntBulletEnv-v0 type: AntBulletEnv-v0 metrics: - type: mean_reward value: 1429.05 +/- 101.46 name: mean_reward verified: false --- # **A2C** Agent playing **AntBulletEnv-v0** This is a trained model of a **A2C** agent playing **AntBulletEnv-v0** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
kiamesdavies/dressup_facial_james_dreambooth_lora_6
kiamesdavies
2023-06-16T15:58:41Z
0
0
diffusers
[ "diffusers", "stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "lora", "base_model:runwayml/stable-diffusion-v1-5", "base_model:adapter:runwayml/stable-diffusion-v1-5", "license:creativeml-openrail-m", "region:us" ]
text-to-image
2023-06-16T15:15:13Z
--- license: creativeml-openrail-m base_model: runwayml/stable-diffusion-v1-5 instance_prompt: a photo of jamesniranye person tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image - diffusers - lora inference: true --- # LoRA DreamBooth - kiamesdavies/dressup_facial_james_dreambooth_lora_6 These are LoRA adaption weights for runwayml/stable-diffusion-v1-5. The weights were trained on a photo of jamesniranye person using [DreamBooth](https://dreambooth.github.io/). You can find some example images in the following. ![img_0](./image_0.png) ![img_1](./image_1.png) ![img_2](./image_2.png) ![img_3](./image_3.png) LoRA for the text encoder was enabled: True.
luischir/robertita-cased-finetuned-squad
luischir
2023-06-16T15:38:43Z
81
0
transformers
[ "transformers", "pytorch", "tensorboard", "roberta", "question-answering", "generated_from_trainer", "license:mit", "endpoints_compatible", "region:us" ]
question-answering
2023-06-16T14:40:42Z
--- license: mit tags: - generated_from_trainer model-index: - name: robertita-cased-finetuned-squad results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # robertita-cased-finetuned-squad This model is a fine-tuned version of [filevich/robertita-cased](https://huggingface.co/filevich/robertita-cased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4716 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.4398 | 1.0 | 1147 | 0.4021 | | 0.2283 | 2.0 | 2294 | 0.4103 | | 0.1291 | 3.0 | 3441 | 0.4716 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3
jiayanli/loan_classifier
jiayanli
2023-06-16T15:32:14Z
3
0
sentence-transformers
[ "sentence-transformers", "pytorch", "mpnet", "setfit", "text-classification", "arxiv:2209.11055", "license:apache-2.0", "region:us" ]
text-classification
2023-06-16T15:31:51Z
--- license: apache-2.0 tags: - setfit - sentence-transformers - text-classification pipeline_tag: text-classification --- # jiayanli/loan_classifier This is a [SetFit model](https://github.com/huggingface/setfit) that can be used for text classification. The model has been trained using an efficient few-shot learning technique that involves: 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning. 2. Training a classification head with features from the fine-tuned Sentence Transformer. ## Usage To use this model for inference, first install the SetFit library: ```bash python -m pip install setfit ``` You can then run inference as follows: ```python from setfit import SetFitModel # Download from Hub and run inference model = SetFitModel.from_pretrained("jiayanli/loan_classifier") # Run inference preds = model(["i loved the spiderman movie!", "pineapple on pizza is the worst 🤮"]) ``` ## BibTeX entry and citation info ```bibtex @article{https://doi.org/10.48550/arxiv.2209.11055, doi = {10.48550/ARXIV.2209.11055}, url = {https://arxiv.org/abs/2209.11055}, author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren}, keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {Efficient Few-Shot Learning Without Prompts}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} } ```
antoninobrillante/gtl-elephant-ext
antoninobrillante
2023-06-16T15:03:26Z
30
0
diffusers
[ "diffusers", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2023-06-16T14:51:26Z
--- license: creativeml-openrail-m tags: - text-to-image - stable-diffusion --- ### gtl-elephant-ext Dreambooth model trained by antoninobrillante with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Sample pictures of this concept:
aga3134/rl_course_vizdoom_health_gathering_supreme
aga3134
2023-06-16T15:00:12Z
0
0
sample-factory
[ "sample-factory", "tensorboard", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T14:59:39Z
--- library_name: sample-factory tags: - deep-reinforcement-learning - reinforcement-learning - sample-factory model-index: - name: APPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: doom_health_gathering_supreme type: doom_health_gathering_supreme metrics: - type: mean_reward value: 11.85 +/- 5.07 name: mean_reward verified: false --- A(n) **APPO** model trained on the **doom_health_gathering_supreme** environment. This model was trained using Sample-Factory 2.0: https://github.com/alex-petrenko/sample-factory. Documentation for how to use Sample-Factory can be found at https://www.samplefactory.dev/ ## Downloading the model After installing Sample-Factory, download the model with: ``` python -m sample_factory.huggingface.load_from_hub -r aga3134/rl_course_vizdoom_health_gathering_supreme ``` ## Using the model To run the model after download, use the `enjoy` script corresponding to this environment: ``` python -m .usr.local.lib.python3.10.dist-packages.ipykernel_launcher --algo=APPO --env=doom_health_gathering_supreme --train_dir=./train_dir --experiment=rl_course_vizdoom_health_gathering_supreme ``` You can also upload models to the Hugging Face Hub using the same script with the `--push_to_hub` flag. See https://www.samplefactory.dev/10-huggingface/huggingface/ for more details ## Training with this model To continue training with this model, use the `train` script corresponding to this environment: ``` python -m .usr.local.lib.python3.10.dist-packages.ipykernel_launcher --algo=APPO --env=doom_health_gathering_supreme --train_dir=./train_dir --experiment=rl_course_vizdoom_health_gathering_supreme --restart_behavior=resume --train_for_env_steps=10000000000 ``` Note, you may have to adjust `--train_for_env_steps` to a suitably high number as the experiment will resume at the number of steps it concluded at.
kejolong/mizuki2.0
kejolong
2023-06-16T14:39:20Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T14:37:05Z
--- license: creativeml-openrail-m ---
AXX1995/gebianv1
AXX1995
2023-06-16T14:32:40Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T14:31:10Z
--- license: creativeml-openrail-m ---
Xuanlong/MUAD_DeepLabmodel
Xuanlong
2023-06-16T14:12:58Z
0
0
null
[ "arxiv:2203.01437", "license:afl-3.0", "region:us" ]
null
2023-06-16T13:11:46Z
--- license: afl-3.0 --- ## DeepLab v3 plus - ResNet101 model trained on MUAD dataset This is a DeepLab v3 plus model with ResNet101 backbone trained on the MUAD dataset. The training is based on PyTorch. MUAD is a synthetic dataset with multiple uncertainties for autonomous driving [[Paper]](https://arxiv.org/abs/2203.01437) [[Website]](https://muad-dataset.github.io/) [[Github]](https://github.com/ENSTA-U2IS/MUAD-Dataset). ### ICCV UNCV 2023 | MUAD challenge MUAD challenge is now on board on the Codalab platform for uncertainty estimation in semantic segmentation. This challenge is hosted in conjunction with the [ICCV 2023](https://iccv2023.thecvf.com/) workshop, [Uncertainty Quantification for Computer Vision (UNCV)](https://uncv2023.github.io/). Go and have a try! 🚀 🚀 🚀 [[Challenge link]](https://codalab.lisn.upsaclay.fr/competitions/8007) ### Reference If you find this work useful for your research, please consider citing our paper: ``` @inproceedings{franchi22bmvc, title = {MUAD: Multiple Uncertainties for Autonomous Driving benchmark for multiple uncertainty types and tasks}, author = {Gianni Franchi and Xuanlong Yu and Andrei Bursuc and Angel Tena and Rémi Kazmierczak and Severine Dubuisson and Emanuel Aldea and David Filliat}, booktitle = {33rd British Machine Vision Conference, {BMVC}}, year = {2022} } ``` ``` @inproceedings{deeplabv3plus2018, title = {Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation}, author = {Liang-Chieh Chen and Yukun Zhu and George Papandreou and Florian Schroff and Hartwig Adam}, booktitle = {ECCV}, year = {2018} } ``` ### Copyright Copyright for MUAD Dataset is owned by Université Paris-Saclay (SATIE Laboratory, Gif-sur-Yvette, FR) and ENSTA Paris (U2IS Laboratory, Palaiseau, FR).
KBLab/bart-base-swedish-cased
KBLab
2023-06-16T14:08:56Z
129
3
transformers
[ "transformers", "pytorch", "safetensors", "bart", "text2text-generation", "sv", "arxiv:1910.13461", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2022-03-02T23:29:04Z
--- language: sv widget: - text: "Jag har ätit en <mask>" --- ## KB-BART A [BART](https://arxiv.org/abs/1910.13461) model trained on a Swedish corpus consisting of 15 billion tokens (about 80GB of text). The model was trained with [Fairseq](https://github.com/pytorch/fairseq), and converted to be compatible with Huggingface. Training code can be found [here](https://github.com/kb-labb/kb_bart). ## Usage ```python from transformers import BartForConditionalGeneration, PreTrainedTokenizerFast, AutoTokenizer model = BartForConditionalGeneration.from_pretrained("KBLab/bart-base-swedish-cased") tok = AutoTokenizer.from_pretrained("KBLab/bart-base-swedish-cased") model.eval() input_ids = tok.encode( "Jag har ätit en utsökt <mask> på restaurang vid <mask> .", return_tensors="pt" ) # Simple greedy search output_ids = model.generate( input_ids, min_length=15, max_length=25, num_beams=1, do_sample=False, ) tok.decode(output_ids[0]) # '</s><s> Jag har ätit en utsökt middag på restaurang vid havet på restaurang vid havet på restaurang vid havet.</s>' # Sampling output_ids = model.generate( input_ids, min_length=15, max_length=20, num_beams=1, do_sample=True, ) tok.decode(output_ids[0]) #'</s><s> Jag har ätit en utsökt god mat som de tagit in på restaurang vid avröjda</s>' # Beam search output_ids = model.generate( input_ids, min_length=15, max_length=25, no_repeat_ngram_size=3, num_beams=8, early_stopping=True, do_sample=True, num_return_sequences=6 ) tok.decode(output_ids[0]) # '</s><s> Jag har ätit en utsökt middag på restaurang vid havet. Jag har varit ute och gått en sväng.</s><pad><pad>' # Diverse beam generation output_ids = model.generate( input_ids, min_length=50, max_length=100, no_repeat_ngram_size=3, num_beams=8, early_stopping=True, do_sample=False, num_return_sequences=6, num_beam_groups=8, diversity_penalty=2.0, ) tok.decode(output_ids[0]) # '</s><s> Jag har ätit en utsökt middag på restaurang vid havet på restaurang. Jag har varit på restaurang i två dagar... Jag..,..!!!.. Så.. Nu.. Hej.. Vi.. Här.</s>' ``` ## Acknowledgements We gratefully acknowledge the HPC RIVR consortium ([www.hpc-rivr.si](https://www.hpc-rivr.si/)) and EuroHPC JU ([eurohpc-ju.europa.eu/](https://eurohpc-ju.europa.eu/)) for funding this research by providing computing resources of the HPC system Vega at the Institute of Information Science ([www.izum.si](https://www.izum.si/)).
VikramDS/my-awesome-setfit-model
VikramDS
2023-06-16T14:07:26Z
4
0
sentence-transformers
[ "sentence-transformers", "pytorch", "mpnet", "setfit", "text-classification", "arxiv:2209.11055", "license:apache-2.0", "region:us" ]
text-classification
2023-06-16T14:06:39Z
--- license: apache-2.0 tags: - setfit - sentence-transformers - text-classification pipeline_tag: text-classification --- # VikramDS/my-awesome-setfit-model This is a [SetFit model](https://github.com/huggingface/setfit) that can be used for text classification. The model has been trained using an efficient few-shot learning technique that involves: 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning. 2. Training a classification head with features from the fine-tuned Sentence Transformer. ## Usage To use this model for inference, first install the SetFit library: ```bash python -m pip install setfit ``` You can then run inference as follows: ```python from setfit import SetFitModel # Download from Hub and run inference model = SetFitModel.from_pretrained("VikramDS/my-awesome-setfit-model") # Run inference preds = model(["i loved the spiderman movie!", "pineapple on pizza is the worst 🤮"]) ``` ## BibTeX entry and citation info ```bibtex @article{https://doi.org/10.48550/arxiv.2209.11055, doi = {10.48550/ARXIV.2209.11055}, url = {https://arxiv.org/abs/2209.11055}, author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren}, keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {Efficient Few-Shot Learning Without Prompts}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} } ```
KHEW/DAIMONDOLLLora
KHEW
2023-06-16T14:06:06Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T14:05:03Z
--- license: creativeml-openrail-m ---
heack/HeackMT5-ZhCleanText1ML
heack
2023-06-16T14:05:27Z
110
11
transformers
[ "transformers", "pytorch", "mt5", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2023-06-15T09:45:13Z
--- pipeline_tag: text2text-generation --- # HeackMT5-ZhCleanText1ML: A Text Cleaning Model for Chinese Texts This model, `heack/HeackMT5-ZhCleanText1ML`, is a fine-tuned mT5 model for Chinese text cleaning tasks. It is designed to remove gibberish, clean up the text, retain original information as much as possible, and does not process large sections of non-Chinese text (such as English text). 此模块,主要解决困扰中国互联网多年的乱码问题,同时借助Transformer大模型,可以对文字进行提炼(很少的情况下以及 模型非常确信的情况下),进行文字清理。你大可以相信此模型,它不会对你的文本进行任意的改动。对于非中文字符的文本, 本模型不做处理。 此模型基于100万行数据进行训练得到,训练结果: | step | epoch | learning_rate | loss | eval_loss |------|-------|------------------------|--------|-------- | 129000 | 3.73 | 1e-05 | 1.714 | 1.706 ## Model Details - Model: mT5 - Language: Chinese (multiple languages supported) ## Usage Here is how you can use this model for text cleaning: ```python from transformers import MT5ForConditionalGeneration, T5Tokenizer model = MT5ForConditionalGeneration.from_pretrained("heack/HeackMT5-ZhCleanText1ML") tokenizer = T5Tokenizer.from_pretrained("heack/HeackMT5-ZhCleanText1ML") text = """ 大众汽车集团在第五届中国国际进口博览会携旗下大众汽车品牌、奥灶液弊胀演蹂穷蹭齿港呛奸怀甫磁洒暮烂犁投迪品牌和保时捷品牌亮相,共展出5款纯电动车 型。其中,大众汽车役络观示惑觉髓品牌展出了ID.家族最新成员——ID.AERO概念车,将于2023年上市;奥迪展出了两款豪华运动纯电动车奥迪RS e-tro???Mission GT和首款“Roadjet 陆地专机”奥迪Q5e-t��������Ʒ�2022��ף��µϽron。到2022年底,奥迪将在中国D��������市场提供7款新能源车型。保时捷则展出了两款纯电动车,其中保时捷Mission R概念车为亚洲首秀。保时捷将进一步在电气化领域持续发力,大量创新技 术萤恒扔剪秆仁忙殃掉雄停遵冒姑只脸玉匣有望应用于未来的量产车中,包括全新的电池组和冷����������却系统等。“自2015年以来,中国在智能汽车领域已逐渐在世界上领先。在自动驾驶领域,没有其他国家的技术创新和实施速度现在能够超越中国。”大众汽车集d 团执行副总裁刘云峰说,他指出,中德双方的务实合作广泛而深入,其中经贸合作发挥了压舱石作鑳藉寲杞�鍨嬬殑涓绘垬鍦轰箣涓�銆用,特别是在掏傻汽车行业。大众汽车集团有关人士介绍,大众正积极主动地推进转型,创新求变,oYFb而中国是大众汽车向电动化和交智能化 转型的主战场之一。除了代表大众迄柑居昧懦汽车电动化攻势的多款纯电车型和创新技术外,大众汽车还在本届进博<script会通过互动形式展示了旗下软件公司CARIAD的最新软件研发成果。按计划,在中国,大众汽车品牌ID.家族浴屋??????????????聂日票绢缀郁硼魏挖两 裙快温屎棠虐惨遇的产品阵容将拓展至纯电中型轿车细分市场。 """ inputs = tokenizer("filter:"+text, return_tensors="pt") outputs = model.generate(inputs.input_ids, max_new_tokens=512) filtered_text = tokenizer.decode(outputs[0], skip_special_tokens=True, num_beams=4, length_penalty=0.8) print(filtered_text) ====================== """ 大众汽车集团在第五届中国国际进口博览会携旗下大众汽车品牌、奥迪品牌和保时捷品牌亮相,共展出5款纯电动车 型。其中,大众汽车品牌展出了ID.家族最新成员——ID.AERO概念车,将于2023年上市;奥迪展出了两款豪华运动纯电动车奥迪RS e-tronMission GT和首款“Roadjet 陆地专机”奥迪Q5e-tron。到2022年底,奥迪将在中国市场提供7款新能源车型。保时捷则展出了两款纯电动车,其中保时捷Mission R概念车为亚洲首秀。保时捷将进一步在电气化领域持续发力,大量创新技 术有望应用于未来的量产车中,包括全新的电池组和冷却系统等。“自2015年以来,中国在智能汽车领域已逐渐在世界上领先。在自动驾驶领域,没有其他国家的技术创新和实施速度现在能够超越中国。”大众汽车集 团执行副总裁刘云峰说,他指出,中德双方的务实合作广泛而深入,其中经贸合作发挥了压舱石作用,特别是在汽车行业。大众汽车集团有关人士介绍,大众正积极主动地推进转型,创新求变,而中国是大众汽车向电动化和交智能化 转型的主战场之一。除了代表大众汽车电动化攻势的多款纯电车型和创新技术外,大众汽车还在本届进博会通过互动形式展示了旗下软件公司CARIAD的最新软件研发成果。按计划,在中国,大众汽车品牌ID.家族的产品阵容将拓展至纯电中型轿车细分市场。 """ ``` ## For long text(more than 512 tokens) ```python from transformers import MT5ForConditionalGeneration, T5Tokenizer def split_text(text, tokenizer, length): chunks = [] chunk = "" for char in text: chunk = chunk + char if len(tokenizer.encode(chunk, truncation=False)) >= length: if char in {'.', '。', ',', ',', '\n'}: chunks.append(chunk) chunk = "" else: for i in range(1, 21): if chunk[-i] in {'.', '。', ',', ',', '\n'}: break else: i = 0 if i == 0: chunks.append(chunk) chunk = "" else: chunks.append(chunk[:-i]) chunk = chunk[-i:] chunks.append(chunk) assert "".join(chunks) == text return chunks def filter_luanma_text(text, model, tokenizer): chunks = split_text(text, tokenizer,500) filter_texts = [] for chunk in chunks: inputs = tokenizer("filter:" + chunk, return_tensors="pt") outputs = model.generate(input_ids=inputs["input_ids"], max_new_tokens=500) filter_text = tokenizer.decode(outputs[0], max_length=500, skip_special_tokens=True, num_beams=4, length_penalty=0.8) filter_texts.append(filter_text) return " ".join(filter_texts) model = MT5ForConditionalGeneration.from_pretrained("heack/HeackMT5-ZhCleanText1ML") tokenizer = T5Tokenizer.from_pretrained("heack/HeackMT5-ZhCleanText1ML") filtered_text = filter_luanma_text("需要df过滤的文=本", model, tokenizer) print(filtered_text) ====================================== """ 需要过滤的文本 """ ``` ## Credits This model is trained and maintained by KongYang from Shanghai Jiao Tong University. For any questions, please reach out to me at my WeChat ID: kongyang. ## License This model is released under the CC BY-NC-SA 4.0 license. ## Citation If you use this model in your research, please cite: ## Citation If you use this model in your research, please cite: ```bibtex @misc{kongyang2023heackmt5ZhCleanText1ML, title={heack/HeackMT5-ZhCleanText1ML: A Large-Scale Multilingual Abstractive Summarization for Chinese Texts}, author={Kong Yang}, year={2023} }
busywhistling/WizardCoder-15B-V1.0_safetensors
busywhistling
2023-06-16T14:01:53Z
12
1
transformers
[ "transformers", "safetensors", "gpt_bigcode", "text-generation", "license:bigcode-openrail-m", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2023-06-16T13:32:32Z
--- license: bigcode-openrail-m ---
studio-ousia/mluke-large
studio-ousia
2023-06-16T13:55:50Z
178
1
transformers
[ "transformers", "pytorch", "luke", "fill-mask", "named entity recognition", "relation classification", "question answering", "multilingual", "ar", "bn", "de", "el", "en", "es", "fi", "fr", "hi", "id", "it", "ja", "ko", "nl", "pl", "pt", "ru", "sv", "sw", "te", "th", "tr", "vi", "zh", "arxiv:2010.01057", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-03-02T23:29:05Z
--- language: - multilingual - ar - bn - de - el - en - es - fi - fr - hi - id - it - ja - ko - nl - pl - pt - ru - sv - sw - te - th - tr - vi - zh thumbnail: https://github.com/studio-ousia/luke/raw/master/resources/luke_logo.png tags: - luke - named entity recognition - relation classification - question answering license: apache-2.0 --- ## mLUKE **mLUKE** (multilingual LUKE) is a multilingual extension of LUKE. Please check the [official repository](https://github.com/studio-ousia/luke) for more details and updates. This is the mLUKE large model with 24 hidden layers, 768 hidden size. The total number of parameters in this model is 868M (561M for the word embeddings and encoder, 307M for the entity embeddings). The model was initialized with the weights of XLM-RoBERTa(large) and trained using December 2020 version of Wikipedia in 24 languages. ## Note When you load the model from `AutoModel.from_pretrained` with the default configuration, you will see the following warning: ``` Some weights of the model checkpoint at studio-ousia/mluke-base-lite were not used when initializing LukeModel: [ 'luke.encoder.layer.0.attention.self.w2e_query.weight', 'luke.encoder.layer.0.attention.self.w2e_query.bias', 'luke.encoder.layer.0.attention.self.e2w_query.weight', 'luke.encoder.layer.0.attention.self.e2w_query.bias', 'luke.encoder.layer.0.attention.self.e2e_query.weight', 'luke.encoder.layer.0.attention.self.e2e_query.bias', ...] ``` These weights are the weights for entity-aware attention (as described in [the LUKE paper](https://arxiv.org/abs/2010.01057)). This is expected because `use_entity_aware_attention` is set to `false` by default, but the pretrained weights contain the weights for it in case you enable `use_entity_aware_attention` and have the weights loaded into the model. ### Citation If you find mLUKE useful for your work, please cite the following paper: ```latex @inproceedings{ri-etal-2022-mluke, title = "m{LUKE}: {T}he Power of Entity Representations in Multilingual Pretrained Language Models", author = "Ri, Ryokan and Yamada, Ikuya and Tsuruoka, Yoshimasa", booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", year = "2022", url = "https://aclanthology.org/2022.acl-long.505", ```
studio-ousia/mluke-large-lite
studio-ousia
2023-06-16T13:55:27Z
1,461
2
transformers
[ "transformers", "pytorch", "luke", "fill-mask", "named entity recognition", "relation classification", "question answering", "multilingual", "ar", "bn", "de", "el", "en", "es", "fi", "fr", "hi", "id", "it", "ja", "ko", "nl", "pl", "pt", "ru", "sv", "sw", "te", "th", "tr", "vi", "zh", "arxiv:2010.01057", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-04-13T10:48:26Z
--- language: - multilingual - ar - bn - de - el - en - es - fi - fr - hi - id - it - ja - ko - nl - pl - pt - ru - sv - sw - te - th - tr - vi - zh thumbnail: https://github.com/studio-ousia/luke/raw/master/resources/luke_logo.png tags: - luke - named entity recognition - relation classification - question answering license: apache-2.0 --- ## mLUKE **mLUKE** (multilingual LUKE) is a multilingual extension of LUKE. Please check the [official repository](https://github.com/studio-ousia/luke) for more details and updates. This is the mLUKE base model with 12 hidden layers, 768 hidden size. The total number of parameters in this model is 561M. The model was initialized with the weights of XLM-RoBERTa(large) and trained using December 2020 version of Wikipedia in 24 languages. This model is a lite-weight version of [studio-ousia/mluke-large](https://huggingface.co/studio-ousia/mluke-large), without Wikipedia entity embeddings but only with special entities such as `[MASK]`. ## Note When you load the model from `AutoModel.from_pretrained` with the default configuration, you will see the following warning: ``` Some weights of the model checkpoint at studio-ousia/mluke-base-lite were not used when initializing LukeModel: [ 'luke.encoder.layer.0.attention.self.w2e_query.weight', 'luke.encoder.layer.0.attention.self.w2e_query.bias', 'luke.encoder.layer.0.attention.self.e2w_query.weight', 'luke.encoder.layer.0.attention.self.e2w_query.bias', 'luke.encoder.layer.0.attention.self.e2e_query.weight', 'luke.encoder.layer.0.attention.self.e2e_query.bias', ...] ``` These weights are the weights for entity-aware attention (as described in [the LUKE paper](https://arxiv.org/abs/2010.01057)). This is expected because `use_entity_aware_attention` is set to `false` by default, but the pretrained weights contain the weights for it in case you enable `use_entity_aware_attention` and have the weights loaded into the model. ### Citation If you find mLUKE useful for your work, please cite the following paper: ```latex @inproceedings{ri-etal-2022-mluke, title = "m{LUKE}: {T}he Power of Entity Representations in Multilingual Pretrained Language Models", author = "Ri, Ryokan and Yamada, Ikuya and Tsuruoka, Yoshimasa", booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", year = "2022", url = "https://aclanthology.org/2022.acl-long.505", ```
studio-ousia/mluke-base
studio-ousia
2023-06-16T13:54:44Z
213
6
transformers
[ "transformers", "pytorch", "luke", "fill-mask", "named entity recognition", "relation classification", "question answering", "multilingual", "ar", "bn", "de", "el", "en", "es", "fi", "fr", "hi", "id", "it", "ja", "ko", "nl", "pl", "pt", "ru", "sv", "sw", "te", "th", "tr", "vi", "zh", "arxiv:2010.01057", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2022-03-02T23:29:05Z
--- language: - multilingual - ar - bn - de - el - en - es - fi - fr - hi - id - it - ja - ko - nl - pl - pt - ru - sv - sw - te - th - tr - vi - zh thumbnail: https://github.com/studio-ousia/luke/raw/master/resources/luke_logo.png tags: - luke - named entity recognition - relation classification - question answering license: apache-2.0 --- ## mLUKE **mLUKE** (multilingual LUKE) is a multilingual extension of LUKE. Please check the [official repository](https://github.com/studio-ousia/luke) for more details and updates. This is the mLUKE base model with 12 hidden layers, 768 hidden size. The total number of parameters in this model is 585M (278M for the word embeddings and encoder, 307M for the entity embeddings). The model was initialized with the weights of XLM-RoBERTa(base) and trained using December 2020 version of Wikipedia in 24 languages. ## Note When you load the model from `AutoModel.from_pretrained` with the default configuration, you will see the following warning: ``` Some weights of the model checkpoint at studio-ousia/mluke-base-lite were not used when initializing LukeModel: [ 'luke.encoder.layer.0.attention.self.w2e_query.weight', 'luke.encoder.layer.0.attention.self.w2e_query.bias', 'luke.encoder.layer.0.attention.self.e2w_query.weight', 'luke.encoder.layer.0.attention.self.e2w_query.bias', 'luke.encoder.layer.0.attention.self.e2e_query.weight', 'luke.encoder.layer.0.attention.self.e2e_query.bias', ...] ``` These weights are the weights for entity-aware attention (as described in [the LUKE paper](https://arxiv.org/abs/2010.01057)). This is expected because `use_entity_aware_attention` is set to `false` by default, but the pretrained weights contain the weights for it in case you enable `use_entity_aware_attention` and have the weights loaded into the model. ### Citation If you find mLUKE useful for your work, please cite the following paper: ```latex @inproceedings{ri-etal-2022-mluke, title = "m{LUKE}: {T}he Power of Entity Representations in Multilingual Pretrained Language Models", author = "Ri, Ryokan and Yamada, Ikuya and Tsuruoka, Yoshimasa", booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", year = "2022", url = "https://aclanthology.org/2022.acl-long.505", ```
vvtq/model_out_4k
vvtq
2023-06-16T13:32:32Z
0
0
diffusers
[ "diffusers", "stable-diffusion", "stable-diffusion-diffusers", "text-to-image", "controlnet", "base_model:runwayml/stable-diffusion-v1-5", "base_model:adapter:runwayml/stable-diffusion-v1-5", "license:creativeml-openrail-m", "region:us" ]
text-to-image
2023-06-16T08:56:16Z
--- license: creativeml-openrail-m base_model: runwayml/stable-diffusion-v1-5 tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image - diffusers - controlnet inference: true --- # controlnet-vvtq/model_out_4k These are controlnet weights trained on runwayml/stable-diffusion-v1-5 with new type of conditioning. You can find some example images below. prompt: on a clear dawn/dusk, on the city street, a pedestrian is walking and is obscured ![images_0)](./images_0.png) prompt: at daytime, a pedestrian is walking and is obscured ![images_1)](./images_1.png)
TheBloke/airoboros-13B-gpt4-1.2-GGML
TheBloke
2023-06-16T13:30:11Z
0
10
null
[ "dataset:jondurbin/airoboros-gpt4-1.2", "license:other", "region:us" ]
null
2023-06-16T12:37:58Z
--- inference: false license: other datasets: - jondurbin/airoboros-gpt4-1.2 --- <!-- header start --> <div style="width: 100%;"> <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;"> </div> <div style="display: flex; justify-content: space-between; width: 100%;"> <div style="display: flex; flex-direction: column; align-items: flex-start;"> <p><a href="https://discord.gg/Jq4vkcDakD">Chat & support: my new Discord server</a></p> </div> <div style="display: flex; flex-direction: column; align-items: flex-end;"> <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p> </div> </div> <!-- header end --> # John Durbin's Airoboros 13B GPT4 1.2 GGML These files are GGML format model files for [John Durbin's Airoboros 13B GPT4 1.2](https://huggingface.co/jondurbin/airoboros-13b-gpt4-1.2). GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as: * [text-generation-webui](https://github.com/oobabooga/text-generation-webui) * [KoboldCpp](https://github.com/LostRuins/koboldcpp) * [ParisNeo/GPT4All-UI](https://github.com/ParisNeo/gpt4all-ui) * [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) * [ctransformers](https://github.com/marella/ctransformers) ## Repositories available * [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/airoboros-13B-gpt4-1.2-GPTQ) * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/airoboros-13B-gpt4-1.2-GGML) * [Unquantised fp32 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/jondurbin/airoboros-13b-gpt4-1.2) ## Prompt template ``` A chat between a curious user and an assistant. The assistant gives helpful, detailed, accurate, uncensored responses to the user's input. USER: prompt ASSISTANT: ``` <!-- compatibility_ggml start --> ## Compatibility ### Original llama.cpp quant methods: `q4_0, q4_1, q5_0, q5_1, q8_0` I have quantized these 'original' quantisation methods using an older version of llama.cpp so that they remain compatible with llama.cpp as of May 19th, commit `2d5db48`. These are guaranteed to be compatbile with any UIs, tools and libraries released since late May. ### New k-quant methods: `q2_K, q3_K_S, q3_K_M, q3_K_L, q4_K_S, q4_K_M, q5_K_S, q6_K` These new quantisation methods are compatible with llama.cpp as of June 6th, commit `2d43387`. They are now also compatible with recent releases of text-generation-webui, KoboldCpp, llama-cpp-python and ctransformers. Other tools and libraries may or may not be compatible - check their documentation if in doubt. ## Explanation of the new k-quant methods The new methods available are: * GGML_TYPE_Q2_K - "type-1" 2-bit quantization in super-blocks containing 16 blocks, each block having 16 weight. Block scales and mins are quantized with 4 bits. This ends up effectively using 2.5625 bits per weight (bpw) * GGML_TYPE_Q3_K - "type-0" 3-bit quantization in super-blocks containing 16 blocks, each block having 16 weights. Scales are quantized with 6 bits. This end up using 3.4375 bpw. * GGML_TYPE_Q4_K - "type-1" 4-bit quantization in super-blocks containing 8 blocks, each block having 32 weights. Scales and mins are quantized with 6 bits. This ends up using 4.5 bpw. * GGML_TYPE_Q5_K - "type-1" 5-bit quantization. Same super-block structure as GGML_TYPE_Q4_K resulting in 5.5 bpw * GGML_TYPE_Q6_K - "type-0" 6-bit quantization. Super-blocks with 16 blocks, each block having 16 weights. Scales are quantized with 8 bits. This ends up using 6.5625 bpw * GGML_TYPE_Q8_K - "type-0" 8-bit quantization. Only used for quantizing intermediate results. The difference to the existing Q8_0 is that the block size is 256. All 2-6 bit dot products are implemented for this quantization type. Refer to the Provided Files table below to see what files use which methods, and how. <!-- compatibility_ggml end --> ## Provided files | Name | Quant method | Bits | Size | Max RAM required | Use case | | ---- | ---- | ---- | ---- | ---- | ----- | | airoboros-13b-gpt4-1.2.ggmlv3.q2_K.bin | q2_K | 2 | 5.51 GB | 8.01 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.vw and feed_forward.w2 tensors, GGML_TYPE_Q2_K for the other tensors. | | airoboros-13b-gpt4-1.2.ggmlv3.q3_K_L.bin | q3_K_L | 3 | 6.93 GB | 9.43 GB | New k-quant method. Uses GGML_TYPE_Q5_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | airoboros-13b-gpt4-1.2.ggmlv3.q3_K_M.bin | q3_K_M | 3 | 6.31 GB | 8.81 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else GGML_TYPE_Q3_K | | airoboros-13b-gpt4-1.2.ggmlv3.q3_K_S.bin | q3_K_S | 3 | 5.66 GB | 8.16 GB | New k-quant method. Uses GGML_TYPE_Q3_K for all tensors | | airoboros-13b-gpt4-1.2.ggmlv3.q4_0.bin | q4_0 | 4 | 7.32 GB | 9.82 GB | Original llama.cpp quant method, 4-bit. | | airoboros-13b-gpt4-1.2.ggmlv3.q4_1.bin | q4_1 | 4 | 8.14 GB | 10.64 GB | Original llama.cpp quant method, 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models. | | airoboros-13b-gpt4-1.2.ggmlv3.q4_K_M.bin | q4_K_M | 4 | 7.87 GB | 10.37 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q4_K | | airoboros-13b-gpt4-1.2.ggmlv3.q4_K_S.bin | q4_K_S | 4 | 7.37 GB | 9.87 GB | New k-quant method. Uses GGML_TYPE_Q4_K for all tensors | | airoboros-13b-gpt4-1.2.ggmlv3.q5_0.bin | q5_0 | 5 | 8.95 GB | 11.45 GB | Original llama.cpp quant method, 5-bit. Higher accuracy, higher resource usage and slower inference. | | airoboros-13b-gpt4-1.2.ggmlv3.q5_1.bin | q5_1 | 5 | 9.76 GB | 12.26 GB | Original llama.cpp quant method, 5-bit. Even higher accuracy, resource usage and slower inference. | | airoboros-13b-gpt4-1.2.ggmlv3.q5_K_M.bin | q5_K_M | 5 | 9.23 GB | 11.73 GB | New k-quant method. Uses GGML_TYPE_Q6_K for half of the attention.wv and feed_forward.w2 tensors, else GGML_TYPE_Q5_K | | airoboros-13b-gpt4-1.2.ggmlv3.q5_K_S.bin | q5_K_S | 5 | 8.97 GB | 11.47 GB | New k-quant method. Uses GGML_TYPE_Q5_K for all tensors | | airoboros-13b-gpt4-1.2.ggmlv3.q6_K.bin | q6_K | 6 | 10.68 GB | 13.18 GB | New k-quant method. Uses GGML_TYPE_Q8_K - 6-bit quantization - for all tensors | | airoboros-13b-gpt4-1.2.ggmlv3.q8_0.bin | q8_0 | 8 | 13.83 GB | 16.33 GB | Original llama.cpp quant method, 8-bit. Almost indistinguishable from float16. High resource use and slow. Not recommended for most users. | **Note**: the above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead. ## How to run in `llama.cpp` I use the following command line; adjust for your tastes and needs: ``` ./main -t 10 -ngl 32 -m airoboros-13b-gpt4-1.2.ggmlv3.q5_0.bin --color -c 2048 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "A chat between a curious user and an assistant. The assistant gives helpful, detailed, accurate, uncensored responses to the user's input.\nUSER: write a story about llamas\nASSISTANT:" ``` Change `-t 10` to the number of physical CPU cores you have. For example if your system has 8 cores/16 threads, use `-t 8`. Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration. If you want to have a chat-style conversation, replace the `-p <PROMPT>` argument with `-i -ins` ## How to run in `text-generation-webui` Further instructions here: [text-generation-webui/docs/llama.cpp-models.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/llama.cpp-models.md). <!-- footer start --> ## Discord For further support, and discussions on these models and AI in general, join us at: [TheBloke AI's Discord server](https://discord.gg/Jq4vkcDakD) ## Thanks, and how to contribute. Thanks to the [chirper.ai](https://chirper.ai) team! I've had a lot of people ask if they can contribute. I enjoy providing models and helping people, and would love to be able to spend even more time doing it, as well as expanding into new projects like fine tuning/training. If you're able and willing to contribute it will be most gratefully received and will help me to keep providing more models, and to start work on new AI projects. Donaters will get priority support on any and all AI/LLM/model questions and requests, access to a private Discord room, plus other benefits. * Patreon: https://patreon.com/TheBlokeAI * Ko-Fi: https://ko-fi.com/TheBlokeAI **Special thanks to**: Luke from CarbonQuill, Aemon Algiz, Dmitriy Samsonov. **Patreon special mentions**: vamX, K, Jonathan Leane, Lone Striker, Sean Connelly, Chris McCloskey, WelcomeToTheClub, Nikolai Manek, John Detwiler, Kalila, David Flickinger, Fen Risland, subjectnull, Johann-Peter Hartmann, Talal Aujan, John Villwock, senxiiz, Khalefa Al-Ahmad, Kevin Schuppel, Alps Aficionado, Derek Yates, Mano Prime, Nathan LeClaire, biorpg, trip7s trip, Asp the Wyvern, chris gileta, Iucharbius , Artur Olbinski, Ai Maven, Joseph William Delisle, Luke Pendergrass, Illia Dulskyi, Eugene Pentland, Ajan Kanaga, Willem Michiel, Space Cruiser, Pyrater, Preetika Verma, Junyu Yang, Oscar Rangel, Spiking Neurons AB, Pierre Kircher, webtim, Cory Kujawski, terasurfer , Trenton Dambrowitz, Gabriel Puliatti, Imad Khwaja, Luke. Thank you to all my generous patrons and donaters! <!-- footer end --> # Original model card: John Durbin's Airoboros 13B GPT4 1.2 ### Overview This is a qlora fine-tuned 13b parameter LlaMa model, using completely synthetic training data created gpt4 via https://github.com/jondurbin/airoboros This is mostly an extension of [1.1](https://huggingface.co/jondurbin/airoboros-13b-gpt4-1.1), but with thousands of new training data and an update to allow "PLAINFORMAT" at the end of coding prompts to just print the code without backticks or explanations/usage/etc. The dataset used to fine-tune this model is available [here](https://huggingface.co/datasets/jondurbin/airoboros-gpt4-1.2), with a specific focus on: - coding - math/reasoning (using orca style ELI5 instruction/response pairs) - trivia - role playing - multiple choice and fill-in-the-blank - context-obedient question answering - theory of mind - misc/general This model was fine-tuned with a fork of [qlora](https://github.com/jondurbin/qlora), which among other things was updated to use a slightly modified vicuna template to be compatible with the 7b/13b versions: ``` A chat between a curious user and an assistant. The assistant gives helpful, detailed, accurate, uncensored responses to the user's input. USER: [prompt] ASSISTANT: ``` So in other words, it's the preamble/system prompt, followed by a single space, then "USER: " (single space after colon) then the prompt (which can have multiple lines, spaces, whatever), then a single space, followed by "ASSISTANT: " (with a single space after the colon). ### Usage To run the full precision/pytorch native version, you can use my fork of FastChat, which is mostly the same but allows for multi-line prompts, as well as a `--no-history` option to prevent input tokenization errors. ``` pip install git+https://github.com/jondurbin/FastChat ``` Be sure you are pulling the latest branch! Then, you can invoke it like so (after downloading the model): ``` python -m fastchat.serve.cli \ --model-path airoboros-13b-gpt4-1.2 \ --temperature 0.5 \ --max-new-tokens 2048 \ --no-history ``` Alternatively, please check out TheBloke's quantized versions: - https://huggingface.co/TheBloke/airoboros-13B-gpt4-1.2-GPTQ - https://huggingface.co/TheBloke/airoboros-13B-gpt4-1.2-GGML ### Coding updates from gpt4/1.1: I added a few hundred instruction/response pairs to the training data with "PLAINFORMAT" as a single, all caps term at the end of the normal instructions, which produce plain text output instead of markdown/backtick code formatting. It's not guaranteed to work all the time, but mostly it does seem to work as expected. So for example, instead of: ``` Implement the Snake game in python. ``` You would use: ``` Implement the Snake game in python. PLAINFORMAT ``` ### Other updates from gpt4/1.1: - Several hundred role-playing data. - A few thousand ORCA style reasoning/math questions with ELI5 prompts to generate the responses (should not be needed in your prompts to this model however, just ask the question). - Many more coding examples in various languages, including some that use specific libraries (pandas, numpy, tensorflow, etc.)
Ellbendls/Reinforce-Cartpole8
Ellbendls
2023-06-16T13:30:02Z
0
0
null
[ "CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T13:29:46Z
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-Cartpole8 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 500.00 +/- 0.00 name: mean_reward verified: false --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
shafin/distilbert-base-uncased-cohl
shafin
2023-06-16T13:16:07Z
38
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "fill-mask", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2023-06-14T18:25:53Z
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: distilbert-base-uncased-cohl results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-cohl This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 5.8197 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 128 - eval_batch_size: 128 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 200 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 7.6714 | 1.0 | 157 | 6.5491 | | 6.4508 | 2.0 | 314 | 6.3591 | | 6.3245 | 3.0 | 471 | 6.2702 | | 6.2262 | 4.0 | 628 | 6.1747 | | 6.1619 | 5.0 | 785 | 6.1199 | | 6.1333 | 6.0 | 942 | 6.0925 | | 6.1038 | 7.0 | 1099 | 6.0610 | | 6.0825 | 8.0 | 1256 | 6.0783 | | 6.0712 | 9.0 | 1413 | 6.0782 | | 6.0594 | 10.0 | 1570 | 6.0546 | | 6.0407 | 11.0 | 1727 | 6.0402 | | 6.036 | 12.0 | 1884 | 6.0381 | | 6.0332 | 13.0 | 2041 | 6.0056 | | 6.0243 | 14.0 | 2198 | 6.0319 | | 6.0156 | 15.0 | 2355 | 6.0127 | | 6.0234 | 16.0 | 2512 | 6.0173 | | 6.0071 | 17.0 | 2669 | 5.9917 | | 6.0029 | 18.0 | 2826 | 5.9979 | | 6.0012 | 19.0 | 2983 | 5.9878 | | 5.9949 | 20.0 | 3140 | 5.9695 | | 5.9894 | 21.0 | 3297 | 5.9852 | | 5.9846 | 22.0 | 3454 | 5.9776 | | 5.9766 | 23.0 | 3611 | 5.9655 | | 5.9787 | 24.0 | 3768 | 5.9602 | | 5.9717 | 25.0 | 3925 | 5.9889 | | 5.9733 | 26.0 | 4082 | 5.9699 | | 5.9655 | 27.0 | 4239 | 5.9611 | | 5.9737 | 28.0 | 4396 | 5.9804 | | 5.9605 | 29.0 | 4553 | 5.9618 | | 5.9623 | 30.0 | 4710 | 5.9489 | | 5.9588 | 31.0 | 4867 | 5.9630 | | 5.9537 | 32.0 | 5024 | 5.9625 | | 5.9536 | 33.0 | 5181 | 5.9692 | | 5.9489 | 34.0 | 5338 | 5.9739 | | 5.9424 | 35.0 | 5495 | 5.9553 | | 5.945 | 36.0 | 5652 | 5.9464 | | 5.9402 | 37.0 | 5809 | 5.9514 | | 5.9376 | 38.0 | 5966 | 5.9398 | | 5.9389 | 39.0 | 6123 | 5.9321 | | 5.9274 | 40.0 | 6280 | 5.9638 | | 5.9324 | 41.0 | 6437 | 5.9382 | | 5.9275 | 42.0 | 6594 | 5.9396 | | 5.9222 | 43.0 | 6751 | 5.9417 | | 5.9282 | 44.0 | 6908 | 5.9344 | | 5.9247 | 45.0 | 7065 | 5.9181 | | 5.9167 | 46.0 | 7222 | 5.9462 | | 5.9099 | 47.0 | 7379 | 5.9378 | | 5.9126 | 48.0 | 7536 | 5.9052 | | 5.9119 | 49.0 | 7693 | 5.9241 | | 5.9116 | 50.0 | 7850 | 5.8920 | | 5.9003 | 51.0 | 8007 | 5.9172 | | 5.8978 | 52.0 | 8164 | 5.9379 | | 5.8994 | 53.0 | 8321 | 5.9163 | | 5.8973 | 54.0 | 8478 | 5.9284 | | 5.8954 | 55.0 | 8635 | 5.9162 | | 5.8959 | 56.0 | 8792 | 5.8985 | | 5.8983 | 57.0 | 8949 | 5.9143 | | 5.8878 | 58.0 | 9106 | 5.9355 | | 5.8909 | 59.0 | 9263 | 5.9024 | | 5.885 | 60.0 | 9420 | 5.9066 | | 5.8861 | 61.0 | 9577 | 5.8989 | | 5.8779 | 62.0 | 9734 | 5.9037 | | 5.8849 | 63.0 | 9891 | 5.8944 | | 5.8819 | 64.0 | 10048 | 5.9009 | | 5.885 | 65.0 | 10205 | 5.9051 | | 5.8747 | 66.0 | 10362 | 5.9144 | | 5.8746 | 67.0 | 10519 | 5.9108 | | 5.8682 | 68.0 | 10676 | 5.8830 | | 5.8763 | 69.0 | 10833 | 5.9133 | | 5.8664 | 70.0 | 10990 | 5.8987 | | 5.8683 | 71.0 | 11147 | 5.8863 | | 5.8675 | 72.0 | 11304 | 5.9088 | | 5.8713 | 73.0 | 11461 | 5.8645 | | 5.8584 | 74.0 | 11618 | 5.9043 | | 5.8657 | 75.0 | 11775 | 5.8824 | | 5.8648 | 76.0 | 11932 | 5.9092 | | 5.8634 | 77.0 | 12089 | 5.9003 | | 5.86 | 78.0 | 12246 | 5.8910 | | 5.8629 | 79.0 | 12403 | 5.8885 | | 5.8505 | 80.0 | 12560 | 5.8681 | | 5.8608 | 81.0 | 12717 | 5.8960 | | 5.8481 | 82.0 | 12874 | 5.9000 | | 5.8495 | 83.0 | 13031 | 5.8935 | | 5.8436 | 84.0 | 13188 | 5.8784 | | 5.8493 | 85.0 | 13345 | 5.8821 | | 5.8507 | 86.0 | 13502 | 5.8831 | | 5.8472 | 87.0 | 13659 | 5.8779 | | 5.8422 | 88.0 | 13816 | 5.8784 | | 5.8412 | 89.0 | 13973 | 5.8630 | | 5.8416 | 90.0 | 14130 | 5.8723 | | 5.842 | 91.0 | 14287 | 5.8794 | | 5.8375 | 92.0 | 14444 | 5.8611 | | 5.8404 | 93.0 | 14601 | 5.8705 | | 5.8451 | 94.0 | 14758 | 5.8883 | | 5.8364 | 95.0 | 14915 | 5.8747 | | 5.8365 | 96.0 | 15072 | 5.8885 | | 5.8277 | 97.0 | 15229 | 5.8667 | | 5.8255 | 98.0 | 15386 | 5.8603 | | 5.8336 | 99.0 | 15543 | 5.8644 | | 5.826 | 100.0 | 15700 | 5.8725 | | 5.8223 | 101.0 | 15857 | 5.8714 | | 5.8415 | 102.0 | 16014 | 5.8773 | | 5.8286 | 103.0 | 16171 | 5.8704 | | 5.8281 | 104.0 | 16328 | 5.8732 | | 5.8246 | 105.0 | 16485 | 5.8582 | | 5.8267 | 106.0 | 16642 | 5.8603 | | 5.8176 | 107.0 | 16799 | 5.8751 | | 5.8214 | 108.0 | 16956 | 5.8774 | | 5.8115 | 109.0 | 17113 | 5.8826 | | 5.8205 | 110.0 | 17270 | 5.8516 | | 5.8136 | 111.0 | 17427 | 5.8743 | | 5.8166 | 112.0 | 17584 | 5.8555 | | 5.8171 | 113.0 | 17741 | 5.8695 | | 5.8176 | 114.0 | 17898 | 5.8531 | | 5.8108 | 115.0 | 18055 | 5.8570 | | 5.808 | 116.0 | 18212 | 5.8552 | | 5.8094 | 117.0 | 18369 | 5.8619 | | 5.8108 | 118.0 | 18526 | 5.8665 | | 5.8064 | 119.0 | 18683 | 5.8851 | | 5.8099 | 120.0 | 18840 | 5.8507 | | 5.8073 | 121.0 | 18997 | 5.8676 | | 5.814 | 122.0 | 19154 | 5.8492 | | 5.8093 | 123.0 | 19311 | 5.8506 | | 5.8135 | 124.0 | 19468 | 5.8668 | | 5.8031 | 125.0 | 19625 | 5.8617 | | 5.801 | 126.0 | 19782 | 5.8626 | | 5.8019 | 127.0 | 19939 | 5.8472 | | 5.8106 | 128.0 | 20096 | 5.8429 | | 5.8013 | 129.0 | 20253 | 5.8668 | | 5.809 | 130.0 | 20410 | 5.8824 | | 5.8 | 131.0 | 20567 | 5.8498 | | 5.8006 | 132.0 | 20724 | 5.8757 | | 5.8008 | 133.0 | 20881 | 5.8397 | | 5.7908 | 134.0 | 21038 | 5.8569 | | 5.7967 | 135.0 | 21195 | 5.8304 | | 5.7908 | 136.0 | 21352 | 5.8265 | | 5.7931 | 137.0 | 21509 | 5.8416 | | 5.7896 | 138.0 | 21666 | 5.8368 | | 5.7904 | 139.0 | 21823 | 5.8608 | | 5.791 | 140.0 | 21980 | 5.8369 | | 5.7887 | 141.0 | 22137 | 5.8705 | | 5.7817 | 142.0 | 22294 | 5.8713 | | 5.787 | 143.0 | 22451 | 5.8488 | | 5.7913 | 144.0 | 22608 | 5.8516 | | 5.7877 | 145.0 | 22765 | 5.8438 | | 5.7905 | 146.0 | 22922 | 5.8595 | | 5.7901 | 147.0 | 23079 | 5.8488 | | 5.7906 | 148.0 | 23236 | 5.8460 | | 5.7806 | 149.0 | 23393 | 5.8294 | | 5.7912 | 150.0 | 23550 | 5.8776 | | 5.7803 | 151.0 | 23707 | 5.8262 | | 5.7821 | 152.0 | 23864 | 5.8729 | | 5.7889 | 153.0 | 24021 | 5.8541 | | 5.783 | 154.0 | 24178 | 5.8542 | | 5.7901 | 155.0 | 24335 | 5.8449 | | 5.7821 | 156.0 | 24492 | 5.8524 | | 5.7868 | 157.0 | 24649 | 5.8675 | | 5.7812 | 158.0 | 24806 | 5.8742 | | 5.7821 | 159.0 | 24963 | 5.8496 | | 5.7851 | 160.0 | 25120 | 5.8463 | | 5.7787 | 161.0 | 25277 | 5.8573 | | 5.7836 | 162.0 | 25434 | 5.8212 | | 5.7786 | 163.0 | 25591 | 5.8683 | | 5.7901 | 164.0 | 25748 | 5.8445 | | 5.7764 | 165.0 | 25905 | 5.8253 | | 5.7793 | 166.0 | 26062 | 5.8443 | | 5.7709 | 167.0 | 26219 | 5.8254 | | 5.7823 | 168.0 | 26376 | 5.8591 | | 5.7753 | 169.0 | 26533 | 5.8154 | | 5.7778 | 170.0 | 26690 | 5.8338 | | 5.7785 | 171.0 | 26847 | 5.8596 | | 5.7658 | 172.0 | 27004 | 5.8644 | | 5.7719 | 173.0 | 27161 | 5.8282 | | 5.781 | 174.0 | 27318 | 5.8451 | | 5.7806 | 175.0 | 27475 | 5.8407 | | 5.7798 | 176.0 | 27632 | 5.8622 | | 5.7772 | 177.0 | 27789 | 5.8445 | | 5.7686 | 178.0 | 27946 | 5.8529 | | 5.7738 | 179.0 | 28103 | 5.8474 | | 5.776 | 180.0 | 28260 | 5.8565 | | 5.7685 | 181.0 | 28417 | 5.8253 | | 5.7659 | 182.0 | 28574 | 5.8449 | | 5.7684 | 183.0 | 28731 | 5.8497 | | 5.7709 | 184.0 | 28888 | 5.8385 | | 5.7631 | 185.0 | 29045 | 5.8131 | | 5.7733 | 186.0 | 29202 | 5.8428 | | 5.7736 | 187.0 | 29359 | 5.8388 | | 5.7704 | 188.0 | 29516 | 5.8519 | | 5.7719 | 189.0 | 29673 | 5.8454 | | 5.7737 | 190.0 | 29830 | 5.8209 | | 5.7667 | 191.0 | 29987 | 5.8681 | | 5.7686 | 192.0 | 30144 | 5.8417 | | 5.7754 | 193.0 | 30301 | 5.8566 | | 5.7743 | 194.0 | 30458 | 5.8510 | | 5.7739 | 195.0 | 30615 | 5.8308 | | 5.7755 | 196.0 | 30772 | 5.8390 | | 5.7702 | 197.0 | 30929 | 5.8320 | | 5.767 | 198.0 | 31086 | 5.8447 | | 5.7691 | 199.0 | 31243 | 5.8465 | | 5.7753 | 200.0 | 31400 | 5.8197 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3
greftly/dqn-SpaceInvadersNoFrameskip-v4
greftly
2023-06-16T13:13:37Z
0
0
stable-baselines3
[ "stable-baselines3", "SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T13:13:02Z
--- library_name: stable-baselines3 tags: - SpaceInvadersNoFrameskip-v4 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: DQN results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: SpaceInvadersNoFrameskip-v4 type: SpaceInvadersNoFrameskip-v4 metrics: - type: mean_reward value: 586.50 +/- 160.62 name: mean_reward verified: false --- # **DQN** Agent playing **SpaceInvadersNoFrameskip-v4** This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib Install the RL Zoo (with SB3 and SB3-Contrib): ```bash pip install rl_zoo3 ``` ``` # Download model and save it into the logs/ folder python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga greftly -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do: ``` python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga greftly -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` ## Training (with the RL Zoo) ``` python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ # Upload the model and generate video (when possible) python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga greftly ``` ## Hyperparameters ```python OrderedDict([('batch_size', 32), ('buffer_size', 100000), ('env_wrapper', ['stable_baselines3.common.atari_wrappers.AtariWrapper']), ('exploration_final_eps', 0.01), ('exploration_fraction', 0.1), ('frame_stack', 4), ('gradient_steps', 1), ('learning_rate', 0.0001), ('learning_starts', 100000), ('n_timesteps', 10000000.0), ('optimize_memory_usage', False), ('policy', 'CnnPolicy'), ('target_update_interval', 1000), ('train_freq', 4), ('normalize', False)]) ``` # Environment Arguments ```python {'render_mode': 'rgb_array'} ```
aguska420/1st_leka
aguska420
2023-06-16T13:07:53Z
0
0
null
[ "license:creativeml-openrail-m", "region:us" ]
null
2023-06-16T13:06:31Z
--- license: creativeml-openrail-m ---
AustinCarthy/OnlyPhishGPT2_suffix_100KP_BFall_fromB_90K_topP_0.75_ratio5
AustinCarthy
2023-06-16T13:00:19Z
0
0
null
[ "tensorboard", "generated_from_trainer", "license:apache-2.0", "region:us" ]
null
2023-06-16T09:32:40Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy - f1 - precision - recall model-index: - name: OnlyPhishGPT2_suffix_100KP_BFall_fromB_90K_topP_0.75_ratio5 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # OnlyPhishGPT2_suffix_100KP_BFall_fromB_90K_topP_0.75_ratio5 This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the Train benign: Fall,Test Benign: Fall, Train phish: Fall, Test phish: Fall, generated url dataset: generated_phish_OnlyPhishGPT2_using_benigh_95KtopP_0.75suffix dataset. It achieves the following results on the evaluation set: - Loss: 0.0247 - Accuracy: 0.9964 - F1: 0.9611 - Precision: 0.9928 - Recall: 0.9314 - Roc Auc Score: 0.9655 - Tpr At Fpr 0.01: 0.8832 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall | Roc Auc Score | Tpr At Fpr 0.01 | |:-------------:|:-----:|:------:|:---------------:|:--------:|:------:|:---------:|:------:|:-------------:|:---------------:| | 0.016 | 1.0 | 35625 | 0.0162 | 0.9964 | 0.9616 | 0.9695 | 0.9538 | 0.9762 | 0.8308 | | 0.0075 | 2.0 | 71250 | 0.0195 | 0.9961 | 0.9586 | 0.9625 | 0.9548 | 0.9765 | 0.772 | | 0.0056 | 3.0 | 106875 | 0.0177 | 0.9968 | 0.9659 | 0.9887 | 0.9442 | 0.9718 | 0.8838 | | 0.0022 | 4.0 | 142500 | 0.0309 | 0.9958 | 0.9543 | 0.9918 | 0.9196 | 0.9596 | 0.865 | | 0.0013 | 5.0 | 178125 | 0.0247 | 0.9964 | 0.9611 | 0.9928 | 0.9314 | 0.9655 | 0.8832 | ### Framework versions - Transformers 4.30.1 - Pytorch 2.0.0+cu118 - Datasets 2.12.0 - Tokenizers 0.13.3
YvesP/Taxi-v3
YvesP
2023-06-16T12:47:27Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T12:47:24Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: Taxi-v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.56 +/- 2.71 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="YvesP/Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
sofia-todeschini/BioBERT-Large-LitCovid-v1.0
sofia-todeschini
2023-06-16T12:42:56Z
106
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-06-16T11:00:34Z
--- tags: - generated_from_trainer metrics: - f1 - accuracy model-index: - name: BioBERT-Large-LitCovid-v1.0 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # BioBERT-Large-LitCovid-v1.0 This model is a fine-tuned version of [dmis-lab/biobert-large-cased-v1.1](https://huggingface.co/dmis-lab/biobert-large-cased-v1.1) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1058 - F1: 0.8993 - Roc Auc: 0.9361 - Accuracy: 0.7969 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | F1 | Roc Auc | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:------:|:-------:|:--------:| | 0.1108 | 1.0 | 6240 | 0.1058 | 0.8993 | 0.9361 | 0.7969 | ### Framework versions - Transformers 4.28.0 - Pytorch 2.0.0 - Datasets 2.1.0 - Tokenizers 0.13.3
gokuls/sa_BERT_48_mnli
gokuls
2023-06-16T12:38:17Z
131
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "en", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-06-16T06:16:12Z
--- language: - en license: apache-2.0 tags: - generated_from_trainer datasets: - glue metrics: - accuracy model-index: - name: sa_BERT_48_mnli results: - task: name: Text Classification type: text-classification dataset: name: GLUE MNLI type: glue config: mnli split: validation_matched args: mnli metrics: - name: Accuracy type: accuracy value: 0.7034174125305126 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # sa_BERT_48_mnli This model is a fine-tuned version of [gokuls/bert_base_48](https://huggingface.co/gokuls/bert_base_48) on the GLUE MNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.7082 - Accuracy: 0.7034 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 4e-05 - train_batch_size: 96 - eval_batch_size: 96 - seed: 10 - distributed_type: multi-GPU - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.9145 | 1.0 | 4091 | 0.8006 | 0.6536 | | 0.7442 | 2.0 | 8182 | 0.7245 | 0.6903 | | 0.6631 | 3.0 | 12273 | 0.7323 | 0.6979 | | 0.5942 | 4.0 | 16364 | 0.7073 | 0.7076 | | 0.5241 | 5.0 | 20455 | 0.7475 | 0.7016 | | 0.4526 | 6.0 | 24546 | 0.8377 | 0.7088 | | 0.3842 | 7.0 | 28637 | 0.8736 | 0.6956 | | 0.3213 | 8.0 | 32728 | 0.9334 | 0.6945 | | 0.2669 | 9.0 | 36819 | 1.0196 | 0.7027 | ### Framework versions - Transformers 4.30.2 - Pytorch 1.14.0a0+410ce96 - Datasets 2.13.0 - Tokenizers 0.13.3
ifti98/taxi-v3
ifti98
2023-06-16T12:33:50Z
0
0
null
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2023-06-16T12:33:46Z
--- tags: - Taxi-v3 - q-learning - reinforcement-learning - custom-implementation model-index: - name: taxi-v3 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: Taxi-v3 type: Taxi-v3 metrics: - type: mean_reward value: 7.56 +/- 2.71 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="ifti98/taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
Falah/Audio_Classification_model
Falah
2023-06-16T12:28:07Z
183
0
transformers
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "audio-classification", "generated_from_trainer", "dataset:minds14", "license:apache-2.0", "endpoints_compatible", "region:us" ]
audio-classification
2023-06-16T12:15:35Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - minds14 metrics: - accuracy model-index: - name: Audio_Classification_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Audio_Classification_model This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the minds14 dataset. It achieves the following results on the evaluation set: - Loss: 2.6498 - Accuracy: 0.0796 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 0.8 | 3 | 2.6440 | 0.0619 | | No log | 1.87 | 7 | 2.6484 | 0.0442 | | 2.6344 | 2.93 | 11 | 2.6489 | 0.0619 | | 2.6344 | 4.0 | 15 | 2.6518 | 0.0619 | | 2.6344 | 4.8 | 18 | 2.6547 | 0.0531 | | 2.6173 | 5.87 | 22 | 2.6512 | 0.0531 | | 2.6173 | 6.93 | 26 | 2.6498 | 0.0796 | | 2.6098 | 8.0 | 30 | 2.6474 | 0.0531 | ### Framework versions - Transformers 4.27.1 - Pytorch 2.0.1+cu118 - Datasets 2.9.0 - Tokenizers 0.13.3
Ioanaaaaaaa/bert-base-uncased-with-preprocess-finetuned-emotion-3-epochs-5e-05-renamed
Ioanaaaaaaa
2023-06-16T12:17:07Z
105
0
transformers
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "dataset:emotion", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2023-06-16T12:06:40Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - emotion metrics: - accuracy - f1 model-index: - name: bert-base-uncased-with-preprocess-finetuned-emotion-3-epochs-5e-05-renamed results: - task: name: Text Classification type: text-classification dataset: name: emotion type: emotion config: split split: validation args: split metrics: - name: Accuracy type: accuracy value: 0.942 - name: F1 type: f1 value: 0.9420914957548232 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-uncased-with-preprocess-finetuned-emotion-3-epochs-5e-05-renamed This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.1279 - Accuracy: 0.942 - F1: 0.9421 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.5277 | 1.0 | 250 | 0.2037 | 0.926 | 0.9257 | | 0.141 | 2.0 | 500 | 0.1352 | 0.9385 | 0.9387 | | 0.0912 | 3.0 | 750 | 0.1279 | 0.942 | 0.9421 | ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu118 - Datasets 2.13.0 - Tokenizers 0.13.3