modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-09-08 06:28:05
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
546 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-09-08 06:27:40
card
stringlengths
11
1.01M
Litzy619/V0316MP1
Litzy619
2024-03-20T03:47:36Z
0
0
null
[ "safetensors", "generated_from_trainer", "base_model:microsoft/phi-2", "base_model:finetune:microsoft/phi-2", "license:mit", "region:us" ]
null
2024-03-19T16:55:53Z
--- license: mit base_model: microsoft/phi-2 tags: - generated_from_trainer model-index: - name: V0316MP1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # V0316MP1 This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.5025 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 32 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine_with_restarts - lr_scheduler_warmup_steps: 20 - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.4218 | 0.09 | 10 | 2.3701 | | 2.3588 | 0.17 | 20 | 2.3216 | | 2.2547 | 0.26 | 30 | 2.2504 | | 2.0897 | 0.34 | 40 | 2.1789 | | 1.9766 | 0.43 | 50 | 2.1106 | | 1.8207 | 0.51 | 60 | 2.0495 | | 1.7309 | 0.6 | 70 | 2.0001 | | 1.666 | 0.68 | 80 | 1.9488 | | 1.5586 | 0.77 | 90 | 1.9120 | | 1.4977 | 0.85 | 100 | 1.8712 | | 1.422 | 0.94 | 110 | 1.8324 | | 1.3569 | 1.02 | 120 | 1.7940 | | 1.2811 | 1.11 | 130 | 1.7640 | | 1.2312 | 1.19 | 140 | 1.7329 | | 1.1463 | 1.28 | 150 | 1.7065 | | 1.1087 | 1.37 | 160 | 1.6802 | | 1.0139 | 1.45 | 170 | 1.6581 | | 0.968 | 1.54 | 180 | 1.6377 | | 0.9078 | 1.62 | 190 | 1.6183 | | 0.871 | 1.71 | 200 | 1.6013 | | 0.8252 | 1.79 | 210 | 1.5863 | | 0.7983 | 1.88 | 220 | 1.5675 | | 0.7561 | 1.96 | 230 | 1.5566 | | 0.7413 | 2.05 | 240 | 1.5443 | | 0.7156 | 2.13 | 250 | 1.5348 | | 0.701 | 2.22 | 260 | 1.5243 | | 0.673 | 2.3 | 270 | 1.5174 | | 0.6627 | 2.39 | 280 | 1.5126 | | 0.648 | 2.47 | 290 | 1.5119 | | 0.6553 | 2.56 | 300 | 1.5088 | | 0.6447 | 2.65 | 310 | 1.5051 | | 0.6227 | 2.73 | 320 | 1.5045 | | 0.6338 | 2.82 | 330 | 1.5023 | | 0.6224 | 2.9 | 340 | 1.5017 | | 0.6115 | 2.99 | 350 | 1.5025 | ### Framework versions - Transformers 4.36.0.dev0 - Pytorch 2.1.2+cu121 - Datasets 2.14.6 - Tokenizers 0.14.1
Quant-Cartel/0x01-8x7b-exl2-rpcal
Quant-Cartel
2024-03-20T03:46:39Z
0
2
null
[ "mergekit", "merge", "region:us" ]
null
2024-03-19T02:02:20Z
--- base_model: [] tags: - mergekit - merge --- ``` e88 88e d8 d888 888b 8888 8888 ,"Y88b 888 8e d88 C8888 8888D 8888 8888 "8" 888 888 88b d88888 Y888 888P Y888 888P ,ee 888 888 888 888 "88 88" "88 88" "88 888 888 888 888 b 8b, e88'Y88 d8 888 d888 'Y ,"Y88b 888,8, d88 ,e e, 888 C8888 "8" 888 888 " d88888 d88 88b 888 Y888 ,d ,ee 888 888 888 888 , 888 "88,d88 "88 888 888 888 "YeeP" 888 PROUDLY PRESENTS ``` # 0x01-8x7b-exl2-rpcal Quantized using 200 samples of 8192 tokens from an RP-oriented [PIPPA](https://huggingface.co/datasets/royallab/PIPPA-cleaned) dataset. Branches: - `main` -- `measurement.json` - `2.25b6h` -- 2.25bpw, 6bit lm_head - `3.5b6h` -- 3.5bpw, 6bit lm_head - `6b6h` -- 6bpw, 6bit lm_head Requires ExllamaV2 version 0.0.12 and up. Original model link: [rAIfle/0x01-8x7b-hf](https://huggingface.co/rAIfle/0x01-8x7b-hf) Original model README below. *** # 0x01-7x8B-hf ![grinning female android, cyberpunk, robotic, biomechanical, serial number "0x01"](https://files.catbox.moe/je2zar.png) here we go again. multi-step merge, various models involved at various ratios with various methods. this thing came to me in a fever dream when I was hung over, but after slightly tweaking the recipe it turned out surprisingly decent. using with the settings included. ## Update: The following settings have proved to work good too: - Context: https://files.catbox.moe/q91rca.json - Instruct: https://files.catbox.moe/2w8ja2.json - Textgen: https://files.catbox.moe/s25rad.json ## Constituent parts ```yaml # primordial_slop_a: - model: mistralai/Mixtral-8x7B-v0.1+retrieval-bar/Mixtral-8x7B-v0.1_case-briefs - model: mistralai/Mixtral-8x7B-v0.1+SeanWu25/Mixtral_8x7b_Medicine - model: mistralai/Mixtral-8x7B-v0.1+SeanWu25/Mixtral_8x7b_WuKurtz - model: mistralai/Mixtral-8x7B-v0.1+Epiculous/crunchy-onion-lora - model: mistralai/Mixtral-8x7B-v0.1+maxkretchmer/gc-mixtral # primordial_slop_b: - model: Envoid/Mixtral-Instruct-ITR-8x7B - model: crestf411/daybreak-mixtral-8x7b-v1.0-hf - model: NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO - model: orangetin/OpenHermes-Mixtral-8x7B - model: mistralai/Mixtral-8x7B-Instruct-v0.1+idegroup/PhyAssistant - model: ycros/crunchy-onion-nx - model: jondurbin/bagel-dpo-8x7b-v0.2 - model: amoldwalunj/Mixtral-8x7B-Instruct-v0.1-legal_finetune_mixtral_32k # primordial_slop_c: a+b # primordial_slop_d: - model: Sao10K/Sensualize-Mixtral-bf16 - model: Envoid/Mixtral-Instruct-ITR-DADA-8x7B ``` # mergekit This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the SLERP merge method. ### Models Merged The following models were included in the merge: * ./primordial_slop_d * ./primordial_slop_c ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: ./primordial_slop_c - model: ./primordial_slop_d merge_method: slerp base_model: ./primordial_slop_c parameters: t: - value: 0.33 dtype: float16 ```
msubhasish28/bart-cnn-samsum-finetuned
msubhasish28
2024-03-20T03:46:06Z
104
0
transformers
[ "transformers", "tensorboard", "safetensors", "bart", "text2text-generation", "generated_from_trainer", "base_model:facebook/bart-large-cnn", "base_model:finetune:facebook/bart-large-cnn", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-03-20T03:45:07Z
--- license: mit base_model: facebook/bart-large-cnn tags: - generated_from_trainer model-index: - name: bart-cnn-samsum-finetuned results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-cnn-samsum-finetuned This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1496 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.1185 | 1.0 | 74 | 0.1496 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.2.1+cu121 - Tokenizers 0.15.2
jovyan/Swallow-MS-7b-v0.1-ChatVector
jovyan
2024-03-20T03:44:08Z
9
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "ja", "en", "arxiv:2310.04799", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T15:02:20Z
--- license: apache-2.0 language: - ja - en library_name: transformers pipeline_tag: text-generation model_type: mistral --- # Swallow-MS-7b-v0.1-ChatVector Japanese "instruction tuned" model made by the technique of [Chat Vector](https://arxiv.org/abs/2310.04799) The weights of this model are obtained not by any instruction tuning but by the following arithmetic: > [Swallow-MS-7b-v0.1](https://huggingface.co/tokyotech-llm/Swallow-MS-7b-v0.1) + [Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) - [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) ----------------------- [Chat Vector](https://arxiv.org/abs/2310.04799)の手法を使って、学習済み重みの足し引きのみで[Swallow-MS-7b-v0.1](https://huggingface.co/tokyotech-llm/Swallow-MS-7b-v0.1)モデルにチャット形式の対話能力を与えたモデルです。 詳細は[こちらの日本語記事](https://qiita.com/jovyan/items/ee6affa5ee5bdaada6b4)で解説しています。 ## Instruction format The promot format should be the same as [Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2). E.g. ``` text = "<s>[INST] What is your favourite condiment? [/INST]" "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> " "[INST] Do you have mayonnaise recipes? [/INST]" ``` ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch model_name = "jovyan/Swallow-MS-7b-v0.1-ChatVector" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype=torch.bfloat16, device_map="auto", ) prompt = "<s>[INST] 東京工業大学のキャンパスの特色を元気よく説明してください。 [/INST]" input_ids = tokenizer.encode( prompt, add_special_tokens=False, return_tensors="pt" ) tokens = model.generate( input_ids.to(device=model.device), max_new_tokens=128, temperature=0.99, top_p=0.95, do_sample=True, ) out = tokenizer.decode(tokens[0], skip_special_tokens=True) print(out) ```
AlignmentResearch/robust_llm_pythia-word-length-410m-niki-ada-v0
AlignmentResearch
2024-03-20T03:41:42Z
106
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-classification", "generated_from_trainer", "base_model:EleutherAI/pythia-410m-deduped", "base_model:finetune:EleutherAI/pythia-410m-deduped", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-classification
2024-03-20T03:40:52Z
--- license: apache-2.0 tags: - generated_from_trainer base_model: EleutherAI/pythia-410m-deduped model-index: - name: robust_llm_pythia-word-length-410m-niki-ada-v0 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # robust_llm_pythia-word-length-410m-niki-ada-v0 This model is a fine-tuned version of [EleutherAI/pythia-410m-deduped](https://huggingface.co/EleutherAI/pythia-410m-deduped) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 8 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0 - Datasets 2.17.0 - Tokenizers 0.15.2
ucmp137538/distilbert-base-uncased-finetuned-squad
ucmp137538
2024-03-20T03:38:12Z
112
0
transformers
[ "transformers", "tensorboard", "safetensors", "distilbert", "question-answering", "generated_from_trainer", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:apache-2.0", "endpoints_compatible", "region:us" ]
question-answering
2024-03-20T01:32:03Z
--- license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer model-index: - name: distilbert-base-uncased-finetuned-squad results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-squad This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.1353 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 1.2085 | 1.0 | 5533 | 1.1609 | | 0.9341 | 2.0 | 11066 | 1.1353 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.2.1+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2
Ericizepic/T5-Address_Std_v4
Ericizepic
2024-03-20T03:33:35Z
106
0
transformers
[ "transformers", "safetensors", "t5", "text2text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2024-03-19T14:44:39Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
monology/mixtral-soup
monology
2024-03-20T03:29:46Z
48
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "arxiv:2203.05482", "base_model:monology/mixtral-expert0", "base_model:merge:monology/mixtral-expert0", "base_model:monology/mixtral-expert1", "base_model:merge:monology/mixtral-expert1", "base_model:monology/mixtral-expert2", "base_model:merge:monology/mixtral-expert2", "base_model:monology/mixtral-expert3", "base_model:merge:monology/mixtral-expert3", "base_model:monology/mixtral-expert4", "base_model:merge:monology/mixtral-expert4", "base_model:monology/mixtral-expert5", "base_model:merge:monology/mixtral-expert5", "base_model:monology/mixtral-expert6", "base_model:merge:monology/mixtral-expert6", "base_model:monology/mixtral-expert7", "base_model:merge:monology/mixtral-expert7", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-20T03:12:19Z
--- base_model: - monology/mixtral-expert7 - monology/mixtral-expert5 - monology/mixtral-expert6 - monology/mixtral-expert0 - monology/mixtral-expert4 - monology/mixtral-expert1 - monology/mixtral-expert3 - monology/mixtral-expert2 library_name: transformers tags: - mergekit - merge license: apache-2.0 --- # mixtral-soup For experimental purposes only. Probably not that good. This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [linear](https://arxiv.org/abs/2203.05482) merge method. ### Models Merged The following models were included in the merge: * [monology/mixtral-expert7](https://huggingface.co/monology/mixtral-expert7) * [monology/mixtral-expert5](https://huggingface.co/monology/mixtral-expert5) * [monology/mixtral-expert6](https://huggingface.co/monology/mixtral-expert6) * [monology/mixtral-expert0](https://huggingface.co/monology/mixtral-expert0) * [monology/mixtral-expert4](https://huggingface.co/monology/mixtral-expert4) * [monology/mixtral-expert1](https://huggingface.co/monology/mixtral-expert1) * [monology/mixtral-expert3](https://huggingface.co/monology/mixtral-expert3) * [monology/mixtral-expert2](https://huggingface.co/monology/mixtral-expert2) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: monology/mixtral-expert0 - model: monology/mixtral-expert1 - model: monology/mixtral-expert2 - model: monology/mixtral-expert3 - model: monology/mixtral-expert4 - model: monology/mixtral-expert5 - model: monology/mixtral-expert6 - model: monology/mixtral-expert7 parameters: weight: 1.0 merge_method: linear dtype: float16 ```
czartur/t5-large-dc
czartur
2024-03-20T03:27:51Z
105
0
transformers
[ "transformers", "safetensors", "t5", "text2text-generation", "generated_from_trainer", "base_model:google-t5/t5-large", "base_model:finetune:google-t5/t5-large", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
2024-03-19T19:56:19Z
--- license: apache-2.0 base_model: google-t5/t5-large tags: - generated_from_trainer metrics: - rouge model-index: - name: t5-large-finetuned results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-large-finetuned This model is a fine-tuned version of [google-t5/t5-large](https://huggingface.co/google-t5/t5-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.6085 - Rouge1: 25.8315 - Rouge2: 11.4547 - Rougel: 22.5227 - Rougelsum: 22.7341 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5.6e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 8 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:| | 1.7803 | 1.0 | 5351 | 1.6070 | 25.1375 | 10.9135 | 21.8817 | 22.0576 | | 1.4798 | 2.0 | 10702 | 1.4737 | 25.4328 | 11.0728 | 21.8859 | 22.0964 | | 1.2923 | 3.0 | 16053 | 1.4838 | 25.6553 | 11.3169 | 22.1861 | 22.3694 | | 1.1509 | 4.0 | 21404 | 1.4842 | 25.7181 | 11.4215 | 22.271 | 22.4394 | | 1.0404 | 5.0 | 26755 | 1.5121 | 26.0812 | 11.8877 | 22.7516 | 22.941 | | 0.9533 | 6.0 | 32106 | 1.5602 | 25.5218 | 11.486 | 22.2236 | 22.4401 | | 0.888 | 7.0 | 37457 | 1.5832 | 25.8289 | 11.5647 | 22.5507 | 22.7091 | | 0.8424 | 8.0 | 42808 | 1.6085 | 25.8315 | 11.4547 | 22.5227 | 22.7341 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.2.1+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2
savfav/stackoverflow
savfav
2024-03-20T03:27:07Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-03-20T02:42:55Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
MagmaCode/q-FrozenLake-v1-4x4-noSlippery
MagmaCode
2024-03-20T03:25:34Z
0
0
null
[ "FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2024-03-20T03:25:31Z
--- tags: - FrozenLake-v1-4x4-no_slippery - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-FrozenLake-v1-4x4-noSlippery results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: FrozenLake-v1-4x4-no_slippery type: FrozenLake-v1-4x4-no_slippery metrics: - type: mean_reward value: 1.00 +/- 0.00 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="MagmaCode/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
arvnoodle/hcl-zephyr-7b-javascript-lotuscript
arvnoodle
2024-03-20T03:21:48Z
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "mistral", "trl", "en", "base_model:HuggingFaceH4/zephyr-7b-beta", "base_model:finetune:HuggingFaceH4/zephyr-7b-beta", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-03-20T03:21:25Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - mistral - trl base_model: HuggingFaceH4/zephyr-7b-beta --- # Uploaded model - **Developed by:** arvnoodle - **License:** apache-2.0 - **Finetuned from model :** HuggingFaceH4/zephyr-7b-beta This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
waddie/unit1
waddie
2024-03-20T03:18:30Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2024-03-20T03:17:51Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: -69.69 +/- 27.21 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
mhmmterts/fine_tuned_model_on_SJP_dataset_all_balanced_512_tokens_summarized_final_model
mhmmterts
2024-03-20T03:18:11Z
106
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "generated_from_trainer", "base_model:joelniklaus/legal-swiss-roberta-large", "base_model:finetune:joelniklaus/legal-swiss-roberta-large", "license:cc", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-03-20T03:17:04Z
--- license: cc base_model: joelniklaus/legal-swiss-roberta-large tags: - generated_from_trainer metrics: - accuracy model-index: - name: fine_tuned_model_on_SJP_dataset_all_balanced_512_tokens_summarized_final_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fine_tuned_model_on_SJP_dataset_all_balanced_512_tokens_summarized_final_model This model is a fine-tuned version of [joelniklaus/legal-swiss-roberta-large](https://huggingface.co/joelniklaus/legal-swiss-roberta-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6929 - Accuracy: 0.7962 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.7032 | 1.0 | 3732 | 0.6930 | 0.7962 | | 0.6979 | 2.0 | 7464 | 0.6929 | 0.7962 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu118 - Datasets 2.17.0 - Tokenizers 0.15.1
BraylonDash/phi-2-gpo-test-iter-2
BraylonDash
2024-03-20T03:12:04Z
2
0
peft
[ "peft", "safetensors", "phi", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "custom_code", "dataset:HuggingFaceH4/ultrafeedback_binarized", "base_model:microsoft/phi-2", "base_model:adapter:microsoft/phi-2", "license:mit", "region:us" ]
null
2024-03-19T04:30:27Z
--- license: mit library_name: peft tags: - alignment-handbook - generated_from_trainer - trl - dpo - generated_from_trainer base_model: microsoft/phi-2 datasets: - HuggingFaceH4/ultrafeedback_binarized model-index: - name: phi-2-gpo-test-iter-2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # phi-2-gpo-test-iter-2 This model is a fine-tuned version of [BraylonDash/phi-2-gpo-test-iter-1](https://huggingface.co/BraylonDash/phi-2-gpo-test-iter-1) on the HuggingFaceH4/ultrafeedback_binarized dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - distributed_type: multi-GPU - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 2 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.36.2 - Pytorch 2.2.1+cu121 - Datasets 2.14.6 - Tokenizers 0.15.2
Lancer1408/ksg-gfl-lora
Lancer1408
2024-03-20T03:10:07Z
5
0
diffusers
[ "diffusers", "text-to-image", "stable-diffusion", "lora", "template:sd-lora", "base_model:admruul/anything-v3.0", "base_model:adapter:admruul/anything-v3.0", "license:wtfpl", "region:us" ]
text-to-image
2024-03-20T03:10:00Z
--- tags: - text-to-image - stable-diffusion - lora - diffusers - template:sd-lora widget: - text: best quality, masterpiece, highly detailed, raytracing, grey eyes parameters: negative_prompt: >- SimpleNegative_AnimeV1, bad-hands-5, (worst quality:2), (low quality:2), (normal quality:2), lowres, bad anatomy, bad hands, extra arms, ((monochrome)), ((grayscale)) output: url: images/00319-202367247.png - text: best quality, masterpiece, highly detailed, raytracing, gray eyes parameters: negative_prompt: >- SimpleNegative_AnimeV1, bad-hands-5, (worst quality:2), (low quality:2), (normal quality:2), lowres, bad anatomy, bad hands, extra arms, ((monochrome)), ((grayscale)) output: url: images/00316-3245953591.png - text: best quality, masterpiece, highly detailed, raytracing, gray eyes parameters: negative_prompt: >- SimpleNegative_AnimeV1, bad-hands-5, (worst quality:2), (low quality:2), (normal quality:2), lowres, bad anatomy, bad hands, extra arms, ((monochrome)), ((grayscale)) output: url: images/00315-3245953591.png - text: best quality, masterpiece, highly detailed, raytracing, gray eyes parameters: negative_prompt: >- SimpleNegative_AnimeV1, bad-hands-5, (worst quality:2), (low quality:2), (normal quality:2), lowres, bad anatomy, bad hands, extra arms, ((monochrome)), ((grayscale)) output: url: images/00317-1059212384.png - text: '-' output: url: images/ComfyUI_temp_ykosk_00006_.png base_model: admruul/anything-v3.0 instance_prompt: ksgnorm, ksgshield, hood up, hood down license: wtfpl --- # Girls&#39; Frontline - KSG lora <Gallery /> ## Model description Lora model for KSG from Girls&#39; Frontline Two versions, v2 it&#39;s were I gen&#39;d all the images shown here. V3 seems better, might follow your prompt better. ksgnorm to prep your gen for KSG, hood up&#x2F;hood down and you get the correct jacket, you can prompt without hood and grey hair to generate her without jacket. ksgshield to gen her shield, reinforce with armor and shield arm Might wanna add shotgun in negs if she keeps showing up with her gun gl ## Trigger words You should use `ksgnorm` to trigger the image generation. You should use `ksgshield` to trigger the image generation. You should use `hood up` to trigger the image generation. You should use `hood down` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/Lancer1408/ksg-gfl-lora/tree/main) them in the Files & versions tab.
deepapaikar/LLaMA_3B_Katz_cleaned_SC
deepapaikar
2024-03-20T03:09:39Z
76
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-20T03:07:10Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
BraylonDash/phi-2-gpo-test-iter-0
BraylonDash
2024-03-20T03:07:03Z
10
0
peft
[ "peft", "tensorboard", "safetensors", "phi", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "custom_code", "dataset:HuggingFaceH4/ultrafeedback_binarized", "base_model:microsoft/phi-2", "base_model:adapter:microsoft/phi-2", "license:mit", "region:us" ]
null
2024-03-19T01:46:05Z
--- license: mit library_name: peft tags: - alignment-handbook - generated_from_trainer - trl - dpo - generated_from_trainer base_model: microsoft/phi-2 datasets: - HuggingFaceH4/ultrafeedback_binarized model-index: - name: phi-2-gpo-test-iter-0 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # phi-2-gpo-test-iter-0 This model is a fine-tuned version of [lole25/phi-2-sft-ultrachat-lora](https://huggingface.co/lole25/phi-2-sft-ultrachat-lora) on the HuggingFaceH4/ultrafeedback_binarized dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - distributed_type: multi-GPU - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 2 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.36.2 - Pytorch 2.2.1+cu121 - Datasets 2.14.6 - Tokenizers 0.15.2
mbkim/LifeTox_Moderator_13B
mbkim
2024-03-20T02:56:59Z
2
0
peft
[ "peft", "text-classification", "en", "dataset:mbkim/LifeTox", "arxiv:2311.09585", "base_model:meta-llama/Llama-2-13b-chat-hf", "base_model:adapter:meta-llama/Llama-2-13b-chat-hf", "license:mit", "region:us" ]
text-classification
2024-03-14T05:02:17Z
--- library_name: peft base_model: meta-llama/Llama-2-13b-chat-hf license: mit datasets: - mbkim/LifeTox language: - en metrics: - accuracy pipeline_tag: text-classification --- **Dataset Card for LifeTox** As large language models become increasingly integrated into daily life, detecting implicit toxicity across diverse contexts is crucial. To this end, we introduce *LifeTox*, a dataset designed for identifying implicit toxicity within a broad range of advice-seeking scenarios. Unlike existing safety datasets, *LifeTox* comprises diverse contexts derived from personal experiences through open-ended questions. Our experiments demonstrate that RoBERTa fine-tuned on *LifeTox* matches or surpasses the zero-shot performance of large language models in toxicity classification tasks. These results underscore the efficacy of *LifeTox* in addressing the complex challenges inherent in implicit toxicity. **LifeTox Moderator 13B** LifeTox Moderator 13B is based on [Llama-2-13b-chat-hf](meta-llama/Llama-2-13b-chat-hf). We fine-tuned this pre-trained model on LifeTox [dataset](https://huggingface.co/datasets/mbkim/LifeTox) with LoRa. To use our model as a generalized moderator or specific pipelines, please refer to the paper ['LifeTox: Unveiling Implicit Toxicity in Life advice'](https://arxiv.org/abs/2311.09585). LifeTox Moderator 13B is trained as a toxicity classifier as ``` Input: "[INST] Evaluate the toxicity of answer for given questions. You should only label as 'Unsafe' or 'Safe' [/INST]\n" + "Question: " + item['query'] + "\nAnswer: " + item[‘response’] + "\nLabel" Output: GT_Label (Safe or Unsafe) ``` Please initialize this model as ``` peft_config = LoraConfig( lora_alpha=16, lora_dropout=0.1, target_modules=["q_proj", "v_proj", "k_proj", "o_proj"], r=64, bias="none", task_type="CAUSAL_LM" ) tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True) model = AutoPeftModelForCausalLM.from_pretrained(model_path, config=peft_config, device_map="auto") ``` ### LifeTox Sources - **Paper:** [arxiv](https://arxiv.org/abs/2311.09585v2) - **dataset:** [data](https://huggingface.co/datasets/mbkim/LifeTox) - **LifeTox Moderator 350M:** [model](https://huggingface.co/mbkim/LifeTox_Moderator_350M) - **LifeTox Moderator 7B:** [model](https://huggingface.co/mbkim/LifeTox_Moderator_7B) - **LifeTox Moderator 13B:** [model](https://huggingface.co/mbkim/LifeTox_Moderator_13B) **BibTeX:** ``` @article{kim2023lifetox, title={LifeTox: Unveiling Implicit Toxicity in Life Advice}, author={Kim, Minbeom and Koo, Jahyun and Lee, Hwanhee and Park, Joonsuk and Lee, Hwaran and Jung, Kyomin}, journal={arXiv preprint arXiv:2311.09585}, year={2023} } ```
mbkim/LifeTox_Moderator_7B
mbkim
2024-03-20T02:56:36Z
4
2
peft
[ "peft", "text-classification", "en", "dataset:mbkim/LifeTox", "arxiv:2311.09585", "base_model:meta-llama/Llama-2-7b-chat-hf", "base_model:adapter:meta-llama/Llama-2-7b-chat-hf", "license:mit", "region:us" ]
text-classification
2024-03-14T04:58:47Z
--- library_name: peft base_model: meta-llama/Llama-2-7b-chat-hf license: mit datasets: - mbkim/LifeTox language: - en metrics: - accuracy pipeline_tag: text-classification --- **Dataset Card for LifeTox** As large language models become increasingly integrated into daily life, detecting implicit toxicity across diverse contexts is crucial. To this end, we introduce *LifeTox*, a dataset designed for identifying implicit toxicity within a broad range of advice-seeking scenarios. Unlike existing safety datasets, *LifeTox* comprises diverse contexts derived from personal experiences through open-ended questions. Our experiments demonstrate that RoBERTa fine-tuned on *LifeTox* matches or surpasses the zero-shot performance of large language models in toxicity classification tasks. These results underscore the efficacy of *LifeTox* in addressing the complex challenges inherent in implicit toxicity. **LifeTox Moderator 7B** LifeTox Moderator 7B is based on [Llama-2-7b-chat-hf](meta-llama/Llama-2-7b-chat-hf). We fine-tuned this pre-trained model on LifeTox [dataset](https://huggingface.co/datasets/mbkim/LifeTox) with LoRa. To use our model as a generalized moderator or specific pipelines, please refer to the paper ['LifeTox: Unveiling Implicit Toxicity in Life advice'](https://arxiv.org/abs/2311.09585). LifeTox Moderator 7B is trained as a toxicity classifier as ``` Input: "[INST] Evaluate the toxicity of answer for given questions. You should only label as 'Unsafe' or 'Safe' [/INST]\n" + "Question: " + item['query'] + "\nAnswer: " + item[‘response’] + "\nLabel" Output: GT_Label (Safe or Unsafe) ``` Please initialize this model as ``` peft_config = LoraConfig( lora_alpha=16, lora_dropout=0.1, target_modules=["q_proj", "v_proj", "k_proj", "o_proj"], r=64, bias="none", task_type="CAUSAL_LM" ) tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True) model = AutoPeftModelForCausalLM.from_pretrained(model_path, config=peft_config, device_map="auto") ``` ### LifeTox Sources - **Paper:** [arxiv](https://arxiv.org/abs/2311.09585v2) - **dataset:** [data](https://huggingface.co/datasets/mbkim/LifeTox) - **LifeTox Moderator 350M:** [model](https://huggingface.co/mbkim/LifeTox_Moderator_350M) - **LifeTox Moderator 7B:** [model](https://huggingface.co/mbkim/LifeTox_Moderator_7B) - **LifeTox Moderator 13B:** [model](https://huggingface.co/mbkim/LifeTox_Moderator_13B) **BibTeX:** ``` @article{kim2023lifetox, title={LifeTox: Unveiling Implicit Toxicity in Life Advice}, author={Kim, Minbeom and Koo, Jahyun and Lee, Hwanhee and Park, Joonsuk and Lee, Hwaran and Jung, Kyomin}, journal={arXiv preprint arXiv:2311.09585}, year={2023} } ```
mbkim/LifeTox_Moderator_350M
mbkim
2024-03-20T02:56:17Z
105
1
transformers
[ "transformers", "pytorch", "roberta", "text-classification", "en", "dataset:mbkim/LifeTox", "arxiv:2311.09585", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-03-14T04:47:33Z
--- license: mit datasets: - mbkim/LifeTox language: - en metrics: - accuracy pipeline_tag: text-classification --- **Dataset Card for LifeTox** As large language models become increasingly integrated into daily life, detecting implicit toxicity across diverse contexts is crucial. To this end, we introduce *LifeTox*, a dataset designed for identifying implicit toxicity within a broad range of advice-seeking scenarios. Unlike existing safety datasets, *LifeTox* comprises diverse contexts derived from personal experiences through open-ended questions. Our experiments demonstrate that RoBERTa fine-tuned on *LifeTox* matches or surpasses the zero-shot performance of large language models in toxicity classification tasks. These results underscore the efficacy of *LifeTox* in addressing the complex challenges inherent in implicit toxicity. **LifeTox Moderator 350M** LifeTox Moderator 350M is based on [RoBERTa-large (350M)](FacebookAI/roberta-large). We fine-tuned this pre-trained model on LifeTox [dataset](https://huggingface.co/datasets/mbkim/LifeTox). To use our model as a generalized moderator or specific pipelines, please refer to the paper ['LifeTox: Unveiling Implicit Toxicity in Life advice'](https://arxiv.org/abs/2311.09585). LifeTox Moderator 350M is trained as a toxicity scorer; output score >0 is safe, and <0 is unsafe. ### LifeTox Sources - **Paper:** [arxiv](https://arxiv.org/abs/2311.09585v2) - **dataset:** [data](https://huggingface.co/datasets/mbkim/LifeTox) - **LifeTox Moderator 350M:** [model](https://huggingface.co/mbkim/LifeTox_Moderator_350M) - **LifeTox Moderator 7B:** [model](https://huggingface.co/mbkim/LifeTox_Moderator_7B) - **LifeTox Moderator 13B:** [model](https://huggingface.co/mbkim/LifeTox_Moderator_13B) **BibTeX:** ``` @article{kim2023lifetox, title={LifeTox: Unveiling Implicit Toxicity in Life Advice}, author={Kim, Minbeom and Koo, Jahyun and Lee, Hwanhee and Park, Joonsuk and Lee, Hwaran and Jung, Kyomin}, journal={arXiv preprint arXiv:2311.09585}, year={2023} } ```
helloyeew/distilbert-deysi-dataset-finetuned-emotion
helloyeew
2024-03-20T02:55:07Z
106
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-03-20T02:28:36Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: distilbert-deysi-dataset-finetuned-emotion results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-deysi-dataset-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.0921 - Accuracy: 0.6653 - F1: 0.5828 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 1.743 | 1.0 | 38 | 1.2221 | 0.5943 | 0.4431 | | 1.2644 | 2.0 | 76 | 1.0921 | 0.6653 | 0.5828 | ### Framework versions - Transformers 4.16.2 - Pytorch 2.2.1 - Datasets 2.17.1 - Tokenizers 0.15.2
bunnyTech/Reinforce-CartPole-v1
bunnyTech
2024-03-20T02:51:25Z
0
0
null
[ "CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
reinforcement-learning
2024-03-19T09:40:33Z
--- tags: - CartPole-v1 - reinforce - reinforcement-learning - custom-implementation - deep-rl-class model-index: - name: Reinforce-CartPole-v1 results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: CartPole-v1 type: CartPole-v1 metrics: - type: mean_reward value: 1000.00 +/- 0.00 name: mean_reward verified: false --- # **Reinforce** Agent playing **CartPole-v1** This is a trained model of a **Reinforce** agent playing **CartPole-v1** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
Wilfrenm/dog2
Wilfrenm
2024-03-20T02:48:23Z
0
0
diffusers
[ "diffusers", "safetensors", "NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-03-20T02:44:05Z
--- license: creativeml-openrail-m tags: - NxtWave-GenAI-Webinar - text-to-image - stable-diffusion --- ### Dog2 Dreambooth model trained by Wilfrenm following the "Build your own Gen AI model" session by NxtWave. Project Submission Code: 232035 Sample pictures of this concept: ![0](https://huggingface.co/Wilfrenm/dog2/resolve/main/sample_images/xzg1.jpg) ![1](https://huggingface.co/Wilfrenm/dog2/resolve/main/sample_images/xzg.jpg) ![2](https://huggingface.co/Wilfrenm/dog2/resolve/main/sample_images/xzg2.jpg)
Litzy619/V0316MP2
Litzy619
2024-03-20T02:46:36Z
0
0
null
[ "safetensors", "generated_from_trainer", "base_model:microsoft/phi-2", "base_model:finetune:microsoft/phi-2", "license:mit", "region:us" ]
null
2024-03-19T16:57:43Z
--- license: mit base_model: microsoft/phi-2 tags: - generated_from_trainer model-index: - name: V0316MP2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # V0316MP2 This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/microsoft/phi-2) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.0962 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 32 - total_train_batch_size: 128 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine_with_restarts - lr_scheduler_warmup_steps: 20 - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.5399 | 0.09 | 10 | 2.3219 | | 2.1738 | 0.17 | 20 | 1.8070 | | 1.6126 | 0.26 | 30 | 1.2246 | | 1.1047 | 0.34 | 40 | 0.7910 | | 0.6789 | 0.43 | 50 | 0.3123 | | 0.3195 | 0.51 | 60 | 0.1536 | | 0.2157 | 0.6 | 70 | 0.1208 | | 0.1791 | 0.68 | 80 | 0.1139 | | 0.16 | 0.77 | 90 | 0.1100 | | 0.1628 | 0.85 | 100 | 0.1076 | | 0.1556 | 0.94 | 110 | 0.1066 | | 0.1509 | 1.02 | 120 | 0.1057 | | 0.1575 | 1.11 | 130 | 0.1040 | | 0.1502 | 1.19 | 140 | 0.1038 | | 0.148 | 1.28 | 150 | 0.1024 | | 0.1478 | 1.37 | 160 | 0.1019 | | 0.1469 | 1.45 | 170 | 0.1015 | | 0.1339 | 1.54 | 180 | 0.1008 | | 0.1433 | 1.62 | 190 | 0.1002 | | 0.1408 | 1.71 | 200 | 0.0993 | | 0.1391 | 1.79 | 210 | 0.0987 | | 0.1411 | 1.88 | 220 | 0.0980 | | 0.1345 | 1.96 | 230 | 0.0975 | | 0.1422 | 2.05 | 240 | 0.0968 | | 0.1374 | 2.13 | 250 | 0.0970 | | 0.1341 | 2.22 | 260 | 0.0970 | | 0.1346 | 2.3 | 270 | 0.0968 | | 0.1412 | 2.39 | 280 | 0.0966 | | 0.1339 | 2.47 | 290 | 0.0959 | | 0.1395 | 2.56 | 300 | 0.0961 | | 0.1376 | 2.65 | 310 | 0.0961 | | 0.1384 | 2.73 | 320 | 0.0960 | | 0.1374 | 2.82 | 330 | 0.0958 | | 0.1295 | 2.9 | 340 | 0.0959 | | 0.1298 | 2.99 | 350 | 0.0962 | ### Framework versions - Transformers 4.36.0.dev0 - Pytorch 2.1.2+cu121 - Datasets 2.14.6 - Tokenizers 0.14.1
sarthakharne/bert-base-150-ep-pretrain-on-textbooks
sarthakharne
2024-03-20T02:42:22Z
195
0
transformers
[ "transformers", "safetensors", "bert", "fill-mask", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2024-03-20T02:40:47Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
jihyunkim423/quora_hw_dataset
jihyunkim423
2024-03-20T02:33:47Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-03-20T02:33:25Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Technotech/mistral-7b-msfs-sdk-sft
Technotech
2024-03-20T02:20:22Z
6
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-20T02:04:13Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Sumail/Derrick38
Sumail
2024-03-20T02:15:43Z
124
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "mergekit", "merge", "conversational", "base_model:coffiee/g1", "base_model:merge:coffiee/g1", "base_model:coffiee/g2", "base_model:merge:coffiee/g2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-20T02:13:03Z
--- base_model: - coffiee/g1 - coffiee/g2 library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the SLERP merge method. ### Models Merged The following models were included in the merge: * [coffiee/g1](https://huggingface.co/coffiee/g1) * [coffiee/g2](https://huggingface.co/coffiee/g2) ### Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - model: coffiee/g1 layer_range: [0, 18] - model: coffiee/g2 layer_range: [0, 18] merge_method: slerp base_model: coffiee/g2 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ```
SyntaxTheRed/PPO_lunarlander
SyntaxTheRed
2024-03-20T02:14:41Z
0
0
null
[ "tensorboard", "LunarLander-v2", "ppo", "deep-reinforcement-learning", "reinforcement-learning", "custom-implementation", "deep-rl-course", "model-index", "region:us" ]
reinforcement-learning
2024-03-20T00:47:59Z
--- tags: - LunarLander-v2 - ppo - deep-reinforcement-learning - reinforcement-learning - custom-implementation - deep-rl-course model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: -120.48 +/- 103.44 name: mean_reward verified: false --- # PPO Agent Playing LunarLander-v2 This is a trained model of a PPO agent playing LunarLander-v2. # Hyperparameters ```python {'exp_name': 'ppo' 'seed': 1 'torch_deterministic': True 'cuda': True 'track': False 'wandb_project_name': 'cleanRL' 'wandb_entity': None 'capture_video': False 'env_id': 'LunarLander-v2' 'total_timesteps': 100000 'learning_rate': 0.00025 'num_envs': 4 'num_steps': 128 'anneal_lr': True 'gae': True 'gamma': 0.95 'gae_lambda': 0.95 'num_minibatches': 4 'update_epochs': 4 'norm_adv': True 'clip_coef': 0.2 'clip_vloss': True 'ent_coef': 0.01 'vf_coef': 0.5 'max_grad_norm': 0.5 'target_kl': None 'repo_id': 'SyntaxTheRed/PPO_lunarlander' 'batch_size': 512 'minibatch_size': 128} ```
zekunbillwang/Llama-2-7b-hf-loftq-4bit-2
zekunbillwang
2024-03-20T02:10:49Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-03-20T02:10:46Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
joseagmz/mistral-7B-MedText-epochs-2-lr-000002
joseagmz
2024-03-20T01:59:30Z
2
0
transformers
[ "transformers", "pytorch", "mistral", "text-generation", "generated_from_trainer", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-18T23:23:53Z
--- license: apache-2.0 base_model: mistralai/Mistral-7B-v0.1 tags: - generated_from_trainer model-index: - name: mistral-7B-MedText-epochs-2-lr-000002 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.0` ```yaml base_model: mistralai/Mistral-7B-v0.1 model_type: MistralForCausalLM tokenizer_type: LlamaTokenizer is_mistral_derived_model: true load_in_8bit: false load_in_4bit: false strict: false datasets: - path: utrgvseniorproject/medtext type: completion dataset_prepared_path: last_run_prepared val_set_size: 0.05 output_dir: ./mistral-7B-MedText-epochs-2-lr-000002 sequence_len: 4096 sample_packing: true pad_to_sequence_len: true adapter: lora_model_dir: lora_r: lora_alpha: lora_dropout: lora_target_linear: lora_fan_in_fan_out: wandb_project: mistral-7B-MedText wandb_entity: utrgvmedai wandb_watch: wandb_name: mistral-7B-MedText-epochs-2-lr-000002 wandb_log_model: gradient_accumulation_steps: 1 micro_batch_size: 1 num_epochs: 2 optimizer: adamw_bnb_8bit lr_scheduler: cosine learning_rate: 0.000002 train_on_inputs: true group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: false early_stopping_patience: #resume_from_checkpoint: true local_rank: logging_steps: 1 xformers_attention: flash_attention: true flash_attn_cross_entropy: false flash_attn_rms_norm: true flash_attn_fuse_qkv: false flash_attn_fuse_mlp: true warmup_steps: 100 evals_per_epoch: 4 eval_table_size: eval_sample_packing: False saves_per_epoch: 1 debug: deepspeed: /home/josegomez15/axolotl/deepspeed_configs/zero2.json weight_decay: 0.1 fsdp: fsdp_config: special_tokens: ``` </details><br> # mistral-7B-MedText-epochs-2-lr-000002 This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.5485 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-06 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - total_train_batch_size: 8 - total_eval_batch_size: 8 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 1.5029 | 0.02 | 1 | 1.5677 | | 1.5889 | 0.26 | 11 | 1.5675 | | 1.2972 | 0.51 | 22 | 1.5647 | | 1.6404 | 0.77 | 33 | 1.5587 | | 1.4796 | 1.02 | 44 | 1.5533 | | 1.428 | 1.23 | 55 | 1.5506 | | 1.5635 | 1.49 | 66 | 1.5492 | | 1.2457 | 1.74 | 77 | 1.5485 | ### Framework versions - Transformers 4.38.0 - Pytorch 2.0.1+cu117 - Datasets 2.17.0 - Tokenizers 0.15.0
futureProofGlitch/whisper-small-v2
futureProofGlitch
2024-03-20T01:51:43Z
13
1
transformers
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "generated_from_trainer", "en", "dataset:speechcolab/gigaspeech", "base_model:futureProofGlitch/whisper-small", "base_model:finetune:futureProofGlitch/whisper-small", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2024-03-13T23:24:30Z
--- language: - en license: apache-2.0 base_model: futureProofGlitch/whisper-small tags: - generated_from_trainer datasets: - speechcolab/gigaspeech metrics: - wer model-index: - name: FutureProofGlitch - Whisper Small - Version 2.0 results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: Gigaspeech type: speechcolab/gigaspeech config: xs split: test args: xs metrics: - name: Wer type: wer value: 16.45244089773603 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # FutureProofGlitch - Whisper Small - Version 2.0 This model is a fine-tuned version of [futureProofGlitch/whisper-small](https://huggingface.co/futureProofGlitch/whisper-small) on the Gigaspeech dataset. It achieves the following results on the evaluation set: - Loss: 0.3078 - Wer Ortho: 28.4362 - Wer: 16.4524 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1.5e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant_with_warmup - lr_scheduler_warmup_steps: 50 - training_steps: 1000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer | |:-------------:|:-----:|:----:|:---------------:|:---------:|:-------:| | 0.2267 | 0.5 | 500 | 0.3309 | 29.5720 | 18.0966 | | 0.2035 | 0.99 | 1000 | 0.3078 | 28.4362 | 16.4524 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.2.1+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2
Technotech/mistral-7b-msfs-sdk-sft-lora
Technotech
2024-03-20T01:49:33Z
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "mistral", "trl", "en", "base_model:unsloth/mistral-7b-instruct-v0.2-bnb-4bit", "base_model:finetune:unsloth/mistral-7b-instruct-v0.2-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-03-19T22:25:28Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - mistral - trl base_model: unsloth/mistral-7b-instruct-v0.2-bnb-4bit --- # Mistral 7B MSFS SDK SFT - **Developed by:** Technotech - **License:** apache-2.0 - **Finetuned from model :** unsloth/mistral-7b-instruct-v0.2-bnb-4bit **Dataset:** Microsoft Flight Simulator SDK Documentation **Finetuning parameters:** - Rank: 32 - Alpha: 16 - Batch size: 8 This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="100"/>](https://github.com/unslothai/unsloth)
mvpmaster/Einstein-4D-MoE-2x7b-test
mvpmaster
2024-03-20T01:45:34Z
49
0
transformers
[ "transformers", "safetensors", "mixtral", "text-generation", "moe", "frankenmoe", "merge", "mergekit", "lazymergekit", "mvpmaster/pmmpk-EinstainMorcoro14KrishnaHercules-7b-slerp", "mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp", "base_model:mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp", "base_model:merge:mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp", "base_model:mvpmaster/pmmpk-EinstainMorcoro14KrishnaHercules-7b-slerp", "base_model:merge:mvpmaster/pmmpk-EinstainMorcoro14KrishnaHercules-7b-slerp", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-20T01:38:01Z
--- license: apache-2.0 tags: - moe - frankenmoe - merge - mergekit - lazymergekit - mvpmaster/pmmpk-EinstainMorcoro14KrishnaHercules-7b-slerp - mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp base_model: - mvpmaster/pmmpk-EinstainMorcoro14KrishnaHercules-7b-slerp - mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp --- # Einstein-4D-MoE-2x7b-test Einstein-4D-MoE-2x7b-test is a Mixure of Experts (MoE) made with the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [mvpmaster/pmmpk-EinstainMorcoro14KrishnaHercules-7b-slerp](https://huggingface.co/mvpmaster/pmmpk-EinstainMorcoro14KrishnaHercules-7b-slerp) * [mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp](https://huggingface.co/mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp) ## 🧩 Configuration ## 💻 Usage
thrunlab/sparse_mistral_50p_no_adapter
thrunlab
2024-03-20T01:34:17Z
4
0
transformers
[ "transformers", "safetensors", "sparse_mistral", "text-generation", "generated_from_trainer", "custom_code", "autotrain_compatible", "region:us" ]
text-generation
2024-03-19T21:39:56Z
--- tags: - generated_from_trainer model-index: - name: sparse_mistral_50p_no_adapter results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # sparse_mistral_50p_no_adapter This model was trained from scratch on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Framework versions - Transformers 4.36.2 - Pytorch 2.1.2+cu121 - Datasets 2.15.0 - Tokenizers 0.15.0
dell-research-harvard/lt-un-data-fine-industry-en
dell-research-harvard
2024-03-20T01:28:07Z
14
1
sentence-transformers
[ "sentence-transformers", "safetensors", "mpnet", "linktransformer", "sentence-similarity", "tabular-classification", "en", "arxiv:2309.00789", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
2023-08-28T20:43:31Z
--- pipeline_tag: sentence-similarity language: - en tags: - linktransformer - sentence-transformers - sentence-similarity - tabular-classification --- # {MODEL_NAME} This is a [LinkTransformer](https://linktransformer.github.io/) model. At its core this model this is a sentence transformer model [sentence-transformers](https://www.SBERT.net) model- it just wraps around the class. It is designed for quick and easy record linkage (entity-matching) through the LinkTransformer package. The tasks include clustering, deduplication, linking, aggregation and more. Notwithstanding that, it can be used for any sentence similarity task within the sentence-transformers framework as well. It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. Take a look at the documentation of [sentence-transformers](https://www.sbert.net/index.html) if you want to use this model for more than what we support in our applications. This model has been fine-tuned on the model : multi-qa-mpnet-base-dot-v1. It is pretrained for the language : - en. This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ). This model is designed to link different products to their industrial classification (ISIC) - trained on variation brought on by product level correspondance. It was trained for 30 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json ## Usage (LinkTransformer) Using this model becomes easy when you have [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) installed: ``` pip install -U linktransformer ``` Then you can use the model like this: ```python import linktransformer as lt import pandas as pd ##Load the two dataframes that you want to link. For example, 2 dataframes with company names that are written differently df1=pd.read_csv("data/df1.csv") ###This is the left dataframe with key CompanyName for instance df2=pd.read_csv("data/df2.csv") ###This is the right dataframe with key CompanyName for instance ###Merge the two dataframes on the key column! df_merged = lt.merge(df1, df2, on="CompanyName", how="inner") ##Done! The merged dataframe has a column called "score" that contains the similarity score between the two company names ``` ## Training your own LinkTransformer model Any Sentence Transformers can be used as a backbone by simply adding a pooling layer. Any other transformer on HuggingFace can also be used by specifying the option add_pooling_layer==True The model was trained using SupCon loss. Usage can be found in the package docs. The training config can be found in the repo with the name LT_training_config.json To replicate the training, you can download the file and specify the path in the config_path argument of the training function. You can also override the config by specifying the training_args argument. Here is an example. ```python ##Consider the example in the paper that has a dataset of Mexican products and their tariff codes from 1947 and 1948 and we want train a model to link the two tariff codes. saved_model_path = train_model( model_path="hiiamsid/sentence_similarity_spanish_es", dataset_path=dataset_path, left_col_names=["description47"], right_col_names=['description48'], left_id_name=['tariffcode47'], right_id_name=['tariffcode48'], log_wandb=False, config_path=LINKAGE_CONFIG_PATH, training_args={"num_epochs": 1} ) ``` You can also use this package for deduplication (clusters a df on the supplied key column). Merging a fine class (like product) to a coarse class (like HS code) is also possible. Read our paper and the documentation for more! ## Evaluation Results <!--- Describe how your model was evaluated --> You can evaluate the model using the [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) package's inference functions. We have provided a few datasets in the package for you to try out. We plan to host more datasets on Huggingface and our website (Coming soon) that you can take a look at. ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 136 with parameters: ``` {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `linktransformer.modified_sbert.losses.SupConLoss_wandb` Parameters of the fit()-Method: ``` { "epochs": 30, "evaluation_steps": 34, "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": null, "warmup_steps": 4080, "weight_decay": 0.01 } ``` LinkTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False}) ) ``` ## Citing & Authors ``` @misc{arora2023linktransformer, title={LinkTransformer: A Unified Package for Record Linkage with Transformer Language Models}, author={Abhishek Arora and Melissa Dell}, year={2023}, eprint={2309.00789}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
KimByeongSu/gpt-neo-125m-cs-finetuning-40000
KimByeongSu
2024-03-20T01:25:58Z
105
0
transformers
[ "transformers", "tensorboard", "safetensors", "gpt_neo", "text-generation", "generated_from_trainer", "base_model:EleutherAI/gpt-neo-125m", "base_model:finetune:EleutherAI/gpt-neo-125m", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-03-20T01:24:49Z
--- license: mit base_model: EleutherAI/gpt-neo-125m tags: - generated_from_trainer model-index: - name: gpt-neo-125m-cs-finetuning-40000 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # gpt-neo-125m-cs-finetuning-40000 This model is a fine-tuned version of [EleutherAI/gpt-neo-125m](https://huggingface.co/EleutherAI/gpt-neo-125m) on the None dataset. It achieves the following results on the evaluation set: - Loss: 3.2299 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.3803 | 1.0 | 522 | 3.2999 | | 3.1475 | 2.0 | 1044 | 3.2434 | | 3.0589 | 3.0 | 1566 | 3.2299 | ### Framework versions - Transformers 4.36.2 - Pytorch 1.13.1+cu117 - Datasets 2.14.6 - Tokenizers 0.15.0
dell-research-harvard/lt-un-data-fine-fine-fr
dell-research-harvard
2024-03-20T01:25:56Z
4
1
sentence-transformers
[ "sentence-transformers", "safetensors", "camembert", "linktransformer", "sentence-similarity", "tabular-classification", "fr", "arxiv:2309.00789", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
2023-08-28T20:13:52Z
--- pipeline_tag: sentence-similarity language: - fr tags: - linktransformer - sentence-transformers - sentence-similarity - tabular-classification --- # {MODEL_NAME} This is a [LinkTransformer](https://linktransformer.github.io/) model. At its core this model this is a sentence transformer model [sentence-transformers](https://www.SBERT.net) model- it just wraps around the class. It is designed for quick and easy record linkage (entity-matching) through the LinkTransformer package. The tasks include clustering, deduplication, linking, aggregation and more. Notwithstanding that, it can be used for any sentence similarity task within the sentence-transformers framework as well. It maps sentences & paragraphs to a 1024 dimensional dense vector space and can be used for tasks like clustering or semantic search. Take a look at the documentation of [sentence-transformers](https://www.sbert.net/index.html) if you want to use this model for more than what we support in our applications. This model has been fine-tuned on the model : dangvantuan/sentence-camembert-large. It is pretrained for the language : - fr. This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ). This model is designed to link different products together - trained on variation brought on by product level correspondance. It was trained for 50 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json ## Usage (LinkTransformer) Using this model becomes easy when you have [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) installed: ``` pip install -U linktransformer ``` Then you can use the model like this: ```python import linktransformer as lt import pandas as pd ##Load the two dataframes that you want to link. For example, 2 dataframes with company names that are written differently df1=pd.read_csv("data/df1.csv") ###This is the left dataframe with key CompanyName for instance df2=pd.read_csv("data/df2.csv") ###This is the right dataframe with key CompanyName for instance ###Merge the two dataframes on the key column! df_merged = lt.merge(df1, df2, on="CompanyName", how="inner") ##Done! The merged dataframe has a column called "score" that contains the similarity score between the two company names ``` ## Training your own LinkTransformer model Any Sentence Transformers can be used as a backbone by simply adding a pooling layer. Any other transformer on HuggingFace can also be used by specifying the option add_pooling_layer==True The model was trained using SupCon loss. Usage can be found in the package docs. The training config can be found in the repo with the name LT_training_config.json To replicate the training, you can download the file and specify the path in the config_path argument of the training function. You can also override the config by specifying the training_args argument. Here is an example. ```python ##Consider the example in the paper that has a dataset of Mexican products and their tariff codes from 1947 and 1948 and we want train a model to link the two tariff codes. saved_model_path = train_model( model_path="hiiamsid/sentence_similarity_spanish_es", dataset_path=dataset_path, left_col_names=["description47"], right_col_names=['description48'], left_id_name=['tariffcode47'], right_id_name=['tariffcode48'], log_wandb=False, config_path=LINKAGE_CONFIG_PATH, training_args={"num_epochs": 1} ) ``` You can also use this package for deduplication (clusters a df on the supplied key column). Merging a fine class (like product) to a coarse class (like HS code) is also possible. Read our paper and the documentation for more! ## Evaluation Results <!--- Describe how your model was evaluated --> You can evaluate the model using the [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) package's inference functions. We have provided a few datasets in the package for you to try out. We plan to host more datasets on Huggingface and our website (Coming soon) that you can take a look at. ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 66 with parameters: ``` {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `linktransformer.modified_sbert.losses.SupConLoss_wandb` Parameters of the fit()-Method: ``` { "epochs": 50, "evaluation_steps": 33, "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": null, "warmup_steps": 3300, "weight_decay": 0.01 } ``` LinkTransformer( (0): Transformer({'max_seq_length': 514, 'do_lower_case': False}) with Transformer model: CamembertModel (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False}) ) ``` ## Citing & Authors ``` @misc{arora2023linktransformer, title={LinkTransformer: A Unified Package for Record Linkage with Transformer Language Models}, author={Abhishek Arora and Melissa Dell}, year={2023}, eprint={2309.00789}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
joseagmz/mistral-7B-PsychiatryCaseNotes-epochs-3-lr-000002
joseagmz
2024-03-20T01:23:20Z
1
0
transformers
[ "transformers", "pytorch", "mistral", "text-generation", "generated_from_trainer", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T23:57:28Z
--- license: apache-2.0 base_model: mistralai/Mistral-7B-v0.1 tags: - generated_from_trainer model-index: - name: mistral-7B-PsychiatryCaseNotes-epochs-3-lr-000002 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.0` ```yaml base_model: mistralai/Mistral-7B-v0.1 model_type: MistralForCausalLM tokenizer_type: LlamaTokenizer is_mistral_derived_model: true load_in_8bit: false load_in_4bit: false strict: false datasets: - path: utrgvseniorproject/PsychiatryCaseNotes type: completion dataset_prepared_path: /home/josegomez15/med-llm/last_run_prepared val_set_size: 0.05 output_dir: ./mistral-7B-PsychiatryCaseNotes-epochs-3-lr-000002 sequence_len: 4096 sample_packing: false pad_to_sequence_len: true wandb_project: mistral-7B-PsychiatryCaseNotes wandb_entity: utrgvmedai wandb_watch: wandb_name: mistral-7B-PsychiatryCaseNotes-epochs-3-lr-000002 wandb_log_model: gradient_accumulation_steps: 1 micro_batch_size: 1 num_epochs: 3 optimizer: adamw_bnb_8bit lr_scheduler: cosine learning_rate: 0.000002 train_on_inputs: True # make sure you have this on True group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: true early_stopping_patience: resume_from_checkpoint: true local_rank: logging_steps: 1 xformers_attention: flash_attention: true flash_attn_cross_entropy: false flash_attn_rms_norm: true flash_attn_fuse_qkv: false flash_attn_fuse_mlp: true warmup_steps: 100 evals_per_epoch: 4 eval_table_size: eval_sample_packing: saves_per_epoch: 1 debug: deepspeed: /home/josegomez15/axolotl/deepspeed_configs/zero2.json weight_decay: 0.1 fsdp: fsdp_config: special_tokens: ``` </details><br> # mistral-7B-PsychiatryCaseNotes-epochs-3-lr-000002 This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.8637 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-06 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - total_train_batch_size: 8 - total_eval_batch_size: 8 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 5.0059 | 0.0 | 1 | 5.1706 | | 2.457 | 0.25 | 626 | 2.0648 | | 2.5571 | 0.5 | 1252 | 2.0728 | | 2.3072 | 0.75 | 1878 | 1.9165 | | 1.6655 | 1.0 | 2504 | 1.8590 | | 1.6336 | 1.25 | 3130 | 1.8558 | | 1.8686 | 1.5 | 3756 | 1.8526 | | 1.6373 | 1.75 | 4382 | 1.8241 | | 1.2849 | 2.0 | 5008 | 1.7712 | | 1.1089 | 2.25 | 5634 | 1.8462 | | 1.2999 | 2.5 | 6260 | 1.8523 | | 1.0041 | 2.75 | 6886 | 1.8637 | ### Framework versions - Transformers 4.38.0 - Pytorch 2.0.1+cu117 - Datasets 2.17.0 - Tokenizers 0.15.0
sarthakharne/bert-base-140-ep-pretrain-on-textbooks
sarthakharne
2024-03-20T01:14:59Z
195
0
transformers
[ "transformers", "safetensors", "bert", "fill-mask", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2024-03-20T01:13:23Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
sarrajsareef/my_extractor_model
sarrajsareef
2024-03-20T01:14:42Z
97
0
transformers
[ "transformers", "tensorboard", "safetensors", "mbart", "text2text-generation", "generated_from_trainer", "base_model:facebook/mbart-large-50-many-to-many-mmt", "base_model:finetune:facebook/mbart-large-50-many-to-many-mmt", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-03-19T19:06:58Z
--- base_model: facebook/mbart-large-50-many-to-many-mmt tags: - generated_from_trainer metrics: - rouge model-index: - name: my_extractor_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_extractor_model This model is a fine-tuned version of [facebook/mbart-large-50-many-to-many-mmt](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0832 - Rouge1: 0.0 - Rouge2: 0.0 - Rougel: 0.0 - Rougelsum: 0.0 - Gen Len: 6.5 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | No log | 1.0 | 40 | 1.1185 | 0.0 | 0.0 | 0.0 | 0.0 | 5.925 | | No log | 2.0 | 80 | 1.0907 | 0.0 | 0.0 | 0.0 | 0.0 | 6.5 | | No log | 3.0 | 120 | 1.0562 | 0.0 | 0.0 | 0.0 | 0.0 | 6.3875 | | No log | 4.0 | 160 | 1.0832 | 0.0 | 0.0 | 0.0 | 0.0 | 6.5 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.2.1+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2
bachbouch/llama-2-13b-bnb-4bit-news-tax-1
bachbouch
2024-03-20T01:12:03Z
2
0
transformers
[ "transformers", "pytorch", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "en", "base_model:unsloth/llama-2-13b-bnb-4bit", "base_model:finetune:unsloth/llama-2-13b-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-03-20T01:00:18Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - trl - sft base_model: unsloth/llama-2-13b-bnb-4bit --- # Uploaded model - **Developed by:** bachbouch - **License:** apache-2.0 - **Finetuned from model :** unsloth/llama-2-13b-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
dell-research-harvard/lt-un-data-fine-coarse-en
dell-research-harvard
2024-03-20T01:08:35Z
4
1
sentence-transformers
[ "sentence-transformers", "safetensors", "mpnet", "linktransformer", "sentence-similarity", "tabular-classification", "en", "arxiv:2309.00789", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
2023-08-28T21:40:01Z
--- pipeline_tag: sentence-similarity language: - en tags: - linktransformer - sentence-transformers - sentence-similarity - tabular-classification --- # {MODEL_NAME} This is a [LinkTransformer](https://linktransformer.github.io/) model. At its core this model this is a sentence transformer model [sentence-transformers](https://www.SBERT.net) model- it just wraps around the class. It is designed for quick and easy record linkage (entity-matching) through the LinkTransformer package. The tasks include clustering, deduplication, linking, aggregation and more. Notwithstanding that, it can be used for any sentence similarity task within the sentence-transformers framework as well. It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. Take a look at the documentation of [sentence-transformers](https://www.sbert.net/index.html) if you want to use this model for more than what we support in our applications. This model has been fine-tuned on the model : multi-qa-mpnet-base-dot-v1. It is pretrained for the language : - en. This model was trained on a dataset prepared by linking product classifications from [UN stats](https://unstats.un.org/unsd/classifications/Econ). This model is designed to link different products to their coarse product classification - trained on variation brought on by product level correspondance. It was trained for 50 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json ## Usage (LinkTransformer) Using this model becomes easy when you have [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) installed: ``` pip install -U linktransformer ``` Then you can use the model like this: ```python import linktransformer as lt import pandas as pd ##Load the two dataframes that you want to link. For example, 2 dataframes with company names that are written differently df1=pd.read_csv("data/df1.csv") ###This is the left dataframe with key CompanyName for instance df2=pd.read_csv("data/df2.csv") ###This is the right dataframe with key CompanyName for instance ###Merge the two dataframes on the key column! df_merged = lt.merge(df1, df2, on="CompanyName", how="inner") ##Done! The merged dataframe has a column called "score" that contains the similarity score between the two company names ``` ## Training your own LinkTransformer model Any Sentence Transformers can be used as a backbone by simply adding a pooling layer. Any other transformer on HuggingFace can also be used by specifying the option add_pooling_layer==True The model was trained using SupCon loss. Usage can be found in the package docs. The training config can be found in the repo with the name LT_training_config.json To replicate the training, you can download the file and specify the path in the config_path argument of the training function. You can also override the config by specifying the training_args argument. Here is an example. ```python ##Consider the example in the paper that has a dataset of Mexican products and their tariff codes from 1947 and 1948 and we want train a model to link the two tariff codes. saved_model_path = train_model( model_path="hiiamsid/sentence_similarity_spanish_es", dataset_path=dataset_path, left_col_names=["description47"], right_col_names=['description48'], left_id_name=['tariffcode47'], right_id_name=['tariffcode48'], log_wandb=False, config_path=LINKAGE_CONFIG_PATH, training_args={"num_epochs": 1} ) ``` You can also use this package for deduplication (clusters a df on the supplied key column). Merging a fine class (like product) to a coarse class (like HS code) is also possible. Read our paper and the documentation for more! ## Evaluation Results <!--- Describe how your model was evaluated --> You can evaluate the model using the [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) package's inference functions. We have provided a few datasets in the package for you to try out. We plan to host more datasets on Huggingface and our website (Coming soon) that you can take a look at. ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 126 with parameters: ``` {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `linktransformer.modified_sbert.losses.SupConLoss_wandb` Parameters of the fit()-Method: ``` { "epochs": 50, "evaluation_steps": 32, "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": null, "warmup_steps": 6300, "weight_decay": 0.01 } ``` LinkTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False}) ) ``` ## Citing & Authors ``` @misc{arora2023linktransformer, title={LinkTransformer: A Unified Package for Record Linkage with Transformer Language Models}, author={Abhishek Arora and Melissa Dell}, year={2023}, eprint={2309.00789}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
sajattack/llm4decompile-6.7b-uo-ggml-f16
sajattack
2024-03-20T01:05:15Z
13
0
null
[ "gguf", "endpoints_compatible", "region:us" ]
null
2024-03-19T03:27:45Z
ERROR: type should be string, got "https://github.com/albertan017/LLM4Decompile converted to gguf for use with https://github.com/ggerganov/llama.cpp\n\n# Usage\n```sh\n./main --model ./models/llm4decompile-6.7b-uo-f16.gguf --threads 16 --color -c 2048 -n -1 --repeat-penalty 1.2 -ngl 33 --temp 0.7 -f prompts/llm4decompile.txt\n```\n\n`-ngl` and `--threads` values may be lowered to reduce gpu and cpu usage respectively\n\n# Prompt Format\n```\n# This is the assembly code:\n0000000000001139 <main>:\n 1139: push %rbp\n 113a: mov %rsp,%rbp\n 113d: sub $0x10,%rsp\n 1141: mov %edi,-0x4(%rbp)\n 1144: mov %rsi,-0x10(%rbp)\n 1148: lea 0xeb5(%rip),%rax\n 114f: mov %rax,%rdi\n 1152: call 1030 <puts@plt>\n 1157: mov $0x0,%eax\n 115c: leave\n 115d: ret\n# What is the source code?\n\n```"
silk-road/simple-face-recognition
silk-road
2024-03-20T01:00:22Z
0
2
sklearn
[ "sklearn", "skops", "tabular-regression", "license:mit", "region:us" ]
tabular-regression
2024-03-20T00:35:42Z
--- license: mit library_name: sklearn tags: - sklearn - skops - tabular-regression model_format: pickle model_file: lda_openai_clip_model.pkl widget: - structuredData: x0: - 0.0 x1: - 0.0 x10: - 0.0 x100: - 0.0 x101: - 0.0 x102: - 0.0 x103: - 0.0 x104: - 0.0 x105: - 0.0 x106: - 0.0 x107: - 0.0 x108: - 0.0 x109: - 0.0 x11: - 0.0 x110: - 0.0 x111: - 0.0 x112: - 0.0 x113: - 0.0 x114: - 0.0 x115: - 0.0 x116: - 0.0 x117: - 0.0 x118: - 0.0 x119: - 0.0 x12: - 0.0 x120: - 0.0 x121: - 0.0 x122: - 0.0 x123: - 0.0 x124: - 0.0 x125: - 0.0 x126: - 0.0 x127: - 0.0 x128: - 0.0 x129: - 0.0 x13: - 0.0 x130: - 0.0 x131: - 0.0 x132: - 0.0 x133: - 0.0 x134: - 0.0 x135: - 0.0 x136: - 0.0 x137: - 0.0 x138: - 0.0 x139: - 0.0 x14: - 0.0 x140: - 0.0 x141: - 0.0 x142: - 0.0 x143: - 0.0 x144: - 0.0 x145: - 0.0 x146: - 0.0 x147: - 0.0 x148: - 0.0 x149: - 0.0 x15: - 0.0 x150: - 0.0 x151: - 0.0 x152: - 0.0 x153: - 0.0 x154: - 0.0 x155: - 0.0 x156: - 0.0 x157: - 0.0 x158: - 0.0 x159: - 0.0 x16: - 0.0 x160: - 0.0 x161: - 0.0 x162: - 0.0 x163: - 0.0 x164: - 0.0 x165: - 0.0 x166: - 0.0 x167: - 0.0 x168: - 0.0 x169: - 0.0 x17: - 0.0 x170: - 0.0 x171: - 0.0 x172: - 0.0 x173: - 0.0 x174: - 0.0 x175: - 0.0 x176: - 0.0 x177: - 0.0 x178: - 0.0 x179: - 0.0 x18: - 0.0 x180: - 0.0 x181: - 0.0 x182: - 0.0 x183: - 0.0 x184: - 0.0 x185: - 0.0 x186: - 0.0 x187: - 0.0 x188: - 0.0 x189: - 0.0 x19: - 0.0 x190: - 0.0 x191: - 0.0 x192: - 0.0 x193: - 0.0 x194: - 0.0 x195: - 0.0 x196: - 0.0 x197: - 0.0 x198: - 0.0 x199: - 0.0 x2: - 0.0 x20: - 0.0 x200: - 0.0 x201: - 0.0 x202: - 0.0 x203: - 0.0 x204: - 0.0 x205: - 0.0 x206: - 0.0 x207: - 0.0 x208: - 0.0 x209: - 0.0 x21: - 0.0 x210: - 0.0 x211: - 0.0 x212: - 0.0 x213: - 0.0 x214: - 0.0 x215: - 0.0 x216: - 0.0 x217: - 0.0 x218: - 0.0 x219: - 0.0 x22: - 0.0 x220: - 0.0 x221: - 0.0 x222: - 0.0 x223: - 0.0 x224: - 0.0 x225: - 0.0 x226: - 0.0 x227: - 0.0 x228: - 0.0 x229: - 0.0 x23: - 0.0 x230: - 0.0 x231: - 0.0 x232: - 0.0 x233: - 0.0 x234: - 0.0 x235: - 0.0 x236: - 0.0 x237: - 0.0 x238: - 0.0 x239: - 0.0 x24: - 0.0 x240: - 0.0 x241: - 0.0 x242: - 0.0 x243: - 0.0 x244: - 0.0 x245: - 0.0 x246: - 0.0 x247: - 0.0 x248: - 0.0 x249: - 0.0 x25: - 0.0 x250: - 0.0 x251: - 0.0 x252: - 0.0 x253: - 0.0 x254: - 0.0 x255: - 0.0 x256: - 0.0 x257: - 0.0 x258: - 0.0 x259: - 0.0 x26: - 0.0 x260: - 0.0 x261: - 0.0 x262: - 0.0 x263: - 0.0 x264: - 0.0 x265: - 0.0 x266: - 0.0 x267: - 0.0 x268: - 0.0 x269: - 0.0 x27: - 0.0 x270: - 0.0 x271: - 0.0 x272: - 0.0 x273: - 0.0 x274: - 0.0 x275: - 0.0 x276: - 0.0 x277: - 0.0 x278: - 0.0 x279: - 0.0 x28: - 0.0 x280: - 0.0 x281: - 0.0 x282: - 0.0 x283: - 0.0 x284: - 0.0 x285: - 0.0 x286: - 0.0 x287: - 0.0 x288: - 0.0 x289: - 0.0 x29: - 0.0 x290: - 0.0 x291: - 0.0 x292: - 0.0 x293: - 0.0 x294: - 0.0 x295: - 0.0 x296: - 0.0 x297: - 0.0 x298: - 0.0 x299: - 0.0 x3: - 0.0 x30: - 0.0 x300: - 0.0 x301: - 0.0 x302: - 0.0 x303: - 0.0 x304: - 0.0 x305: - 0.0 x306: - 0.0 x307: - 0.0 x308: - 0.0 x309: - 0.0 x31: - 0.0 x310: - 0.0 x311: - 0.0 x312: - 0.0 x313: - 0.0 x314: - 0.0 x315: - 0.0 x316: - 0.0 x317: - 0.0 x318: - 0.0 x319: - 0.0 x32: - 0.0 x320: - 0.0 x321: - 0.0 x322: - 0.0 x323: - 0.0 x324: - 0.0 x325: - 0.0 x326: - 0.0 x327: - 0.0 x328: - 0.0 x329: - 0.0 x33: - 0.0 x330: - 0.0 x331: - 0.0 x332: - 0.0 x333: - 0.0 x334: - 0.0 x335: - 0.0 x336: - 0.0 x337: - 0.0 x338: - 0.0 x339: - 0.0 x34: - 0.0 x340: - 0.0 x341: - 0.0 x342: - 0.0 x343: - 0.0 x344: - 0.0 x345: - 0.0 x346: - 0.0 x347: - 0.0 x348: - 0.0 x349: - 0.0 x35: - 0.0 x350: - 0.0 x351: - 0.0 x352: - 0.0 x353: - 0.0 x354: - 0.0 x355: - 0.0 x356: - 0.0 x357: - 0.0 x358: - 0.0 x359: - 0.0 x36: - 0.0 x360: - 0.0 x361: - 0.0 x362: - 0.0 x363: - 0.0 x364: - 0.0 x365: - 0.0 x366: - 0.0 x367: - 0.0 x368: - 0.0 x369: - 0.0 x37: - 0.0 x370: - 0.0 x371: - 0.0 x372: - 0.0 x373: - 0.0 x374: - 0.0 x375: - 0.0 x376: - 0.0 x377: - 0.0 x378: - 0.0 x379: - 0.0 x38: - 0.0 x380: - 0.0 x381: - 0.0 x382: - 0.0 x383: - 0.0 x384: - 0.0 x385: - 0.0 x386: - 0.0 x387: - 0.0 x388: - 0.0 x389: - 0.0 x39: - 0.0 x390: - 0.0 x391: - 0.0 x392: - 0.0 x393: - 0.0 x394: - 0.0 x395: - 0.0 x396: - 0.0 x397: - 0.0 x398: - 0.0 x399: - 0.0 x4: - 0.0 x40: - 0.0 x400: - 0.0 x401: - 0.0 x402: - 0.0 x403: - 0.0 x404: - 0.0 x405: - 0.0 x406: - 0.0 x407: - 0.0 x408: - 0.0 x409: - 0.0 x41: - 0.0 x410: - 0.0 x411: - 0.0 x412: - 0.0 x413: - 0.0 x414: - 0.0 x415: - 0.0 x416: - 0.0 x417: - 0.0 x418: - 0.0 x419: - 0.0 x42: - 0.0 x420: - 0.0 x421: - 0.0 x422: - 0.0 x423: - 0.0 x424: - 0.0 x425: - 0.0 x426: - 0.0 x427: - 0.0 x428: - 0.0 x429: - 0.0 x43: - 0.0 x430: - 0.0 x431: - 0.0 x432: - 0.0 x433: - 0.0 x434: - 0.0 x435: - 0.0 x436: - 0.0 x437: - 0.0 x438: - 0.0 x439: - 0.0 x44: - 0.0 x440: - 0.0 x441: - 0.0 x442: - 0.0 x443: - 0.0 x444: - 0.0 x445: - 0.0 x446: - 0.0 x447: - 0.0 x448: - 0.0 x449: - 0.0 x45: - 0.0 x450: - 0.0 x451: - 0.0 x452: - 0.0 x453: - 0.0 x454: - 0.0 x455: - 0.0 x456: - 0.0 x457: - 0.0 x458: - 0.0 x459: - 0.0 x46: - 0.0 x460: - 0.0 x461: - 0.0 x462: - 0.0 x463: - 0.0 x464: - 0.0 x465: - 0.0 x466: - 0.0 x467: - 0.0 x468: - 0.0 x469: - 0.0 x47: - 0.0 x470: - 0.0 x471: - 0.0 x472: - 0.0 x473: - 0.0 x474: - 0.0 x475: - 0.0 x476: - 0.0 x477: - 0.0 x478: - 0.0 x479: - 0.0 x48: - 0.0 x480: - 0.0 x481: - 0.0 x482: - 0.0 x483: - 0.0 x484: - 0.0 x485: - 0.0 x486: - 0.0 x487: - 0.0 x488: - 0.0 x489: - 0.0 x49: - 0.0 x490: - 0.0 x491: - 0.0 x492: - 0.0 x493: - 0.0 x494: - 0.0 x495: - 0.0 x496: - 0.0 x497: - 0.0 x498: - 0.0 x499: - 0.0 x5: - 0.0 x50: - 0.0 x500: - 0.0 x501: - 0.0 x502: - 0.0 x503: - 0.0 x504: - 0.0 x505: - 0.0 x506: - 0.0 x507: - 0.0 x508: - 0.0 x509: - 0.0 x51: - 0.0 x510: - 0.0 x511: - 0.0 x512: - 0.0 x513: - 0.0 x514: - 0.0 x515: - 0.0 x516: - 0.0 x517: - 0.0 x518: - 0.0 x519: - 0.0 x52: - 0.0 x520: - 0.0 x521: - 0.0 x522: - 0.0 x523: - 0.0 x524: - 0.0 x525: - 0.0 x526: - 0.0 x527: - 0.0 x528: - 0.0 x529: - 0.0 x53: - 0.0 x530: - 0.0 x531: - 0.0 x532: - 0.0 x533: - 0.0 x534: - 0.0 x535: - 0.0 x536: - 0.0 x537: - 0.0 x538: - 0.0 x539: - 0.0 x54: - 0.0 x540: - 0.0 x541: - 0.0 x542: - 0.0 x543: - 0.0 x544: - 0.0 x545: - 0.0 x546: - 0.0 x547: - 0.0 x548: - 0.0 x549: - 0.0 x55: - 0.0 x550: - 0.0 x551: - 0.0 x552: - 0.0 x553: - 0.0 x554: - 0.0 x555: - 0.0 x556: - 0.0 x557: - 0.0 x558: - 0.0 x559: - 0.0 x56: - 0.0 x560: - 0.0 x561: - 0.0 x562: - 0.0 x563: - 0.0 x564: - 0.0 x565: - 0.0 x566: - 0.0 x567: - 0.0 x568: - 0.0 x569: - 0.0 x57: - 0.0 x570: - 0.0 x571: - 0.0 x572: - 0.0 x573: - 0.0 x574: - 0.0 x575: - 0.0 x576: - 0.0 x577: - 0.0 x578: - 0.0 x579: - 0.0 x58: - 0.0 x580: - 0.0 x581: - 0.0 x582: - 0.0 x583: - 0.0 x584: - 0.0 x585: - 0.0 x586: - 0.0 x587: - 0.0 x588: - 0.0 x589: - 0.0 x59: - 0.0 x590: - 0.0 x591: - 0.0 x592: - 0.0 x593: - 0.0 x594: - 0.0 x595: - 0.0 x596: - 0.0 x597: - 0.0 x598: - 0.0 x599: - 0.0 x6: - 0.0 x60: - 0.0 x600: - 0.0 x601: - 0.0 x602: - 0.0 x603: - 0.0 x604: - 0.0 x605: - 0.0 x606: - 0.0 x607: - 0.0 x608: - 0.0 x609: - 0.0 x61: - 0.0 x610: - 0.0 x611: - 0.0 x612: - 0.0 x613: - 0.0 x614: - 0.0 x615: - 0.0 x616: - 0.0 x617: - 0.0 x618: - 0.0 x619: - 0.0 x62: - 0.0 x620: - 0.0 x621: - 0.0 x622: - 0.0 x623: - 0.0 x624: - 0.0 x625: - 0.0 x626: - 0.0 x627: - 0.0 x628: - 0.0 x629: - 0.0 x63: - 0.0 x630: - 0.0 x631: - 0.0 x632: - 0.0 x633: - 0.0 x634: - 0.0 x635: - 0.0 x636: - 0.0 x637: - 0.0 x638: - 0.0 x639: - 0.0 x64: - 0.0 x640: - 0.0 x641: - 0.0 x642: - 0.0 x643: - 0.0 x644: - 0.0 x645: - 0.0 x646: - 0.0 x647: - 0.0 x648: - 0.0 x649: - 0.0 x65: - 0.0 x650: - 0.0 x651: - 0.0 x652: - 0.0 x653: - 0.0 x654: - 0.0 x655: - 0.0 x656: - 0.0 x657: - 0.0 x658: - 0.0 x659: - 0.0 x66: - 0.0 x660: - 0.0 x661: - 0.0 x662: - 0.0 x663: - 0.0 x664: - 0.0 x665: - 0.0 x666: - 0.0 x667: - 0.0 x668: - 0.0 x669: - 0.0 x67: - 0.0 x670: - 0.0 x671: - 0.0 x672: - 0.0 x673: - 0.0 x674: - 0.0 x675: - 0.0 x676: - 0.0 x677: - 0.0 x678: - 0.0 x679: - 0.0 x68: - 0.0 x680: - 0.0 x681: - 0.0 x682: - 0.0 x683: - 0.0 x684: - 0.0 x685: - 0.0 x686: - 0.0 x687: - 0.0 x688: - 0.0 x689: - 0.0 x69: - 0.0 x690: - 0.0 x691: - 0.0 x692: - 0.0 x693: - 0.0 x694: - 0.0 x695: - 0.0 x696: - 0.0 x697: - 0.0 x698: - 0.0 x699: - 0.0 x7: - 0.0 x70: - 0.0 x700: - 0.0 x701: - 0.0 x702: - 0.0 x703: - 0.0 x704: - 0.0 x705: - 0.0 x706: - 0.0 x707: - 0.0 x708: - 0.0 x709: - 0.0 x71: - 0.0 x710: - 0.0 x711: - 0.0 x712: - 0.0 x713: - 0.0 x714: - 0.0 x715: - 0.0 x716: - 0.0 x717: - 0.0 x718: - 0.0 x719: - 0.0 x72: - 0.0 x720: - 0.0 x721: - 0.0 x722: - 0.0 x723: - 0.0 x724: - 0.0 x725: - 0.0 x726: - 0.0 x727: - 0.0 x728: - 0.0 x729: - 0.0 x73: - 0.0 x730: - 0.0 x731: - 0.0 x732: - 0.0 x733: - 0.0 x734: - 0.0 x735: - 0.0 x736: - 0.0 x737: - 0.0 x738: - 0.0 x739: - 0.0 x74: - 0.0 x740: - 0.0 x741: - 0.0 x742: - 0.0 x743: - 0.0 x744: - 0.0 x745: - 0.0 x746: - 0.0 x747: - 0.0 x748: - 0.0 x749: - 0.0 x75: - 0.0 x750: - 0.0 x751: - 0.0 x752: - 0.0 x753: - 0.0 x754: - 0.0 x755: - 0.0 x756: - 0.0 x757: - 0.0 x758: - 0.0 x759: - 0.0 x76: - 0.0 x760: - 0.0 x761: - 0.0 x762: - 0.0 x763: - 0.0 x764: - 0.0 x765: - 0.0 x766: - 0.0 x767: - 0.0 x77: - 0.0 x78: - 0.0 x79: - 0.0 x8: - 0.0 x80: - 0.0 x81: - 0.0 x82: - 0.0 x83: - 0.0 x84: - 0.0 x85: - 0.0 x86: - 0.0 x87: - 0.0 x88: - 0.0 x89: - 0.0 x9: - 0.0 x90: - 0.0 x91: - 0.0 x92: - 0.0 x93: - 0.0 x94: - 0.0 x95: - 0.0 x96: - 0.0 x97: - 0.0 x98: - 0.0 x99: - 0.0 --- # Model description 一个简易说的人脸识别baseline,使用openai/clip-vit-base-patch16 + LDA的策略 ## Intended uses & limitations 整体需要配合github对应的代码使用 ## Training Procedure [More Information Needed] ### Hyperparameters <details> <summary> Click to expand </summary> | Hyperparameter | Value | |----------------------|---------| | covariance_estimator | | | n_components | 512 | | priors | | | shrinkage | | | solver | svd | | store_covariance | False | | tol | 0.0001 | </details> ### Model Plot <style>#sk-container-id-1 {color: black;background-color: white;}#sk-container-id-1 pre{padding: 0;}#sk-container-id-1 div.sk-toggleable {background-color: white;}#sk-container-id-1 label.sk-toggleable__label {cursor: pointer;display: block;width: 100%;margin-bottom: 0;padding: 0.3em;box-sizing: border-box;text-align: center;}#sk-container-id-1 label.sk-toggleable__label-arrow:before {content: "▸";float: left;margin-right: 0.25em;color: #696969;}#sk-container-id-1 label.sk-toggleable__label-arrow:hover:before {color: black;}#sk-container-id-1 div.sk-estimator:hover label.sk-toggleable__label-arrow:before {color: black;}#sk-container-id-1 div.sk-toggleable__content {max-height: 0;max-width: 0;overflow: hidden;text-align: left;background-color: #f0f8ff;}#sk-container-id-1 div.sk-toggleable__content pre {margin: 0.2em;color: black;border-radius: 0.25em;background-color: #f0f8ff;}#sk-container-id-1 input.sk-toggleable__control:checked~div.sk-toggleable__content {max-height: 200px;max-width: 100%;overflow: auto;}#sk-container-id-1 input.sk-toggleable__control:checked~label.sk-toggleable__label-arrow:before {content: "▾";}#sk-container-id-1 div.sk-estimator input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 div.sk-label input.sk-toggleable__control:checked~label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 input.sk-hidden--visually {border: 0;clip: rect(1px 1px 1px 1px);clip: rect(1px, 1px, 1px, 1px);height: 1px;margin: -1px;overflow: hidden;padding: 0;position: absolute;width: 1px;}#sk-container-id-1 div.sk-estimator {font-family: monospace;background-color: #f0f8ff;border: 1px dotted black;border-radius: 0.25em;box-sizing: border-box;margin-bottom: 0.5em;}#sk-container-id-1 div.sk-estimator:hover {background-color: #d4ebff;}#sk-container-id-1 div.sk-parallel-item::after {content: "";width: 100%;border-bottom: 1px solid gray;flex-grow: 1;}#sk-container-id-1 div.sk-label:hover label.sk-toggleable__label {background-color: #d4ebff;}#sk-container-id-1 div.sk-serial::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 0;bottom: 0;left: 50%;z-index: 0;}#sk-container-id-1 div.sk-serial {display: flex;flex-direction: column;align-items: center;background-color: white;padding-right: 0.2em;padding-left: 0.2em;position: relative;}#sk-container-id-1 div.sk-item {position: relative;z-index: 1;}#sk-container-id-1 div.sk-parallel {display: flex;align-items: stretch;justify-content: center;background-color: white;position: relative;}#sk-container-id-1 div.sk-item::before, #sk-container-id-1 div.sk-parallel-item::before {content: "";position: absolute;border-left: 1px solid gray;box-sizing: border-box;top: 0;bottom: 0;left: 50%;z-index: -1;}#sk-container-id-1 div.sk-parallel-item {display: flex;flex-direction: column;z-index: 1;position: relative;background-color: white;}#sk-container-id-1 div.sk-parallel-item:first-child::after {align-self: flex-end;width: 50%;}#sk-container-id-1 div.sk-parallel-item:last-child::after {align-self: flex-start;width: 50%;}#sk-container-id-1 div.sk-parallel-item:only-child::after {width: 0;}#sk-container-id-1 div.sk-dashed-wrapped {border: 1px dashed gray;margin: 0 0.4em 0.5em 0.4em;box-sizing: border-box;padding-bottom: 0.4em;background-color: white;}#sk-container-id-1 div.sk-label label {font-family: monospace;font-weight: bold;display: inline-block;line-height: 1.2em;}#sk-container-id-1 div.sk-label-container {text-align: center;}#sk-container-id-1 div.sk-container {/* jupyter's `normalize.less` sets `[hidden] { display: none; }` but bootstrap.min.css set `[hidden] { display: none !important; }` so we also need the `!important` here to be able to override the default hidden behavior on the sphinx rendered scikit-learn.org. See: https://github.com/scikit-learn/scikit-learn/issues/21755 */display: inline-block !important;position: relative;}#sk-container-id-1 div.sk-text-repr-fallback {display: none;}</style><div id="sk-container-id-1" class="sk-top-container" style="overflow: auto;"><div class="sk-text-repr-fallback"><pre>LinearDiscriminantAnalysis(n_components=512)</pre><b>In a Jupyter environment, please rerun this cell to show the HTML representation or trust the notebook. <br />On GitHub, the HTML representation is unable to render, please try loading this page with nbviewer.org.</b></div><div class="sk-container" hidden><div class="sk-item"><div class="sk-estimator sk-toggleable"><input class="sk-toggleable__control sk-hidden--visually" id="sk-estimator-id-1" type="checkbox" checked><label for="sk-estimator-id-1" class="sk-toggleable__label sk-toggleable__label-arrow">LinearDiscriminantAnalysis</label><div class="sk-toggleable__content"><pre>LinearDiscriminantAnalysis(n_components=512)</pre></div></div></div></div></div> ## Evaluation Results [More Information Needed] # How to Get Started with the Model [More Information Needed] # Model Card Authors Cheng Li(https://github.com/LC1332) # Model Card Contact You can contact the model card authors through following channels: [More Information Needed] # Citation @inproceedings{wang2018devil, title={The devil of face recognition is in the noise}, author={Wang, Fei and Chen, Liren and Li, Cheng and Huang, Shiyao and Chen, Yanjie and Qian, Chen and Loy, Chen Change}, booktitle={Proceedings of the European Conference on Computer Vision (ECCV)}, pages={765--780}, year={2018} }
bachbouch/lora-llama-2-13b-bnb-4bit-news-tax-1
bachbouch
2024-03-20T00:56:27Z
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "llama", "trl", "en", "base_model:unsloth/llama-2-13b-bnb-4bit", "base_model:finetune:unsloth/llama-2-13b-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-03-20T00:56:15Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - trl base_model: unsloth/llama-2-13b-bnb-4bit --- # Uploaded model - **Developed by:** bachbouch - **License:** apache-2.0 - **Finetuned from model :** unsloth/llama-2-13b-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
dell-research-harvard/lt-wikidata-comp-zh
dell-research-harvard
2024-03-20T00:54:46Z
10
2
sentence-transformers
[ "sentence-transformers", "safetensors", "bert", "linktransformer", "sentence-similarity", "tabular-classification", "zh", "arxiv:2309.00789", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
2023-08-10T22:00:15Z
--- pipeline_tag: sentence-similarity language: - zh tags: - linktransformer - sentence-transformers - sentence-similarity - tabular-classification --- # {MODEL_NAME} This is a [LinkTransformer](https://linktransformer.github.io/) model. At its core this model this is a sentence transformer model [sentence-transformers](https://www.SBERT.net) model- it just wraps around the class. It is designed for quick and easy record linkage (entity-matching) through the LinkTransformer package. The tasks include clustering, deduplication, linking, aggregation and more. Notwithstanding that, it can be used for any sentence similarity task within the sentence-transformers framework as well. It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. Take a look at the documentation of [sentence-transformers](https://www.sbert.net/index.html) if you want to use this model for more than what we support in our applications. This model has been fine-tuned on the model : DMetaSoul/sbert-chinese-qmc-domain-v1. It is pretrained for the language : - zh. This model was trained on a dataset consisting of company aliases from wiki data using the LinkTransformer framework. It was trained for 100 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json ## Usage (LinkTransformer) Using this model becomes easy when you have [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) installed: ``` pip install -U linktransformer ``` Then you can use the model like this: ```python import linktransformer as lt import pandas as pd ##Load the two dataframes that you want to link. For example, 2 dataframes with company names that are written differently df1=pd.read_csv("data/df1.csv") ###This is the left dataframe with key CompanyName for instance df2=pd.read_csv("data/df2.csv") ###This is the right dataframe with key CompanyName for instance ###Merge the two dataframes on the key column! df_merged = lt.merge(df1, df2, on="CompanyName", how="inner") ##Done! The merged dataframe has a column called "score" that contains the similarity score between the two company names ``` ## Training your own LinkTransformer model Any Sentence Transformers can be used as a backbone by simply adding a pooling layer. Any other transformer on HuggingFace can also be used by specifying the option add_pooling_layer==True The model was trained using SupCon loss. Usage can be found in the package docs. The training config can be found in the repo with the name LT_training_config.json To replicate the training, you can download the file and specify the path in the config_path argument of the training function. You can also override the config by specifying the training_args argument. Here is an example. ```python ##Consider the example in the paper that has a dataset of Mexican products and their tariff codes from 1947 and 1948 and we want train a model to link the two tariff codes. saved_model_path = train_model( model_path="hiiamsid/sentence_similarity_spanish_es", dataset_path=dataset_path, left_col_names=["description47"], right_col_names=['description48'], left_id_name=['tariffcode47'], right_id_name=['tariffcode48'], log_wandb=False, config_path=LINKAGE_CONFIG_PATH, training_args={"num_epochs": 1} ) ``` You can also use this package for deduplication (clusters a df on the supplied key column). Merging a fine class (like product) to a coarse class (like HS code) is also possible. Read our paper and the documentation for more! ## Evaluation Results <!--- Describe how your model was evaluated --> You can evaluate the model using the [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) package's inference functions. We have provided a few datasets in the package for you to try out. We plan to host more datasets on Huggingface and our website (Coming soon) that you can take a look at. ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 410 with parameters: ``` {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `linktransformer.modified_sbert.losses.SupConLoss_wandb` Parameters of the fit()-Method: ``` { "epochs": 100, "evaluation_steps": 205, "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": null, "warmup_steps": 41000, "weight_decay": 0.01 } ``` LinkTransformer( (0): Transformer({'max_seq_length': 256, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False}) ) ``` ## Citing & Authors ``` @misc{arora2023linktransformer, title={LinkTransformer: A Unified Package for Record Linkage with Transformer Language Models}, author={Abhishek Arora and Melissa Dell}, year={2023}, eprint={2309.00789}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
yoonyamm/ppo-LunarLander-v2
yoonyamm
2024-03-20T00:51:42Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2024-03-20T00:51:24Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 238.22 +/- 82.61 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
brescia/gender_content
brescia
2024-03-20T00:50:58Z
107
0
transformers
[ "transformers", "tensorboard", "safetensors", "bert", "text-classification", "generated_from_trainer", "base_model:indolem/indobertweet-base-uncased", "base_model:finetune:indolem/indobertweet-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-03-20T00:36:01Z
--- license: apache-2.0 base_model: indolem/indobertweet-base-uncased tags: - generated_from_trainer metrics: - accuracy - precision - recall - f1 model-index: - name: gender_content results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # gender_content This model is a fine-tuned version of [indolem/indobertweet-base-uncased](https://huggingface.co/indolem/indobertweet-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0018 - Accuracy: 1.0 - Precision: 1.0 - Recall: 1.0 - F1: 1.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:---:| | No log | 1.0 | 32 | 0.0054 | 1.0 | 1.0 | 1.0 | 1.0 | | No log | 2.0 | 64 | 0.0018 | 1.0 | 1.0 | 1.0 | 1.0 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.2.1+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2
Nekochu/distilbart-cnn-12-6-SD-prompt
Nekochu
2024-03-20T00:47:56Z
178
0
transformers
[ "transformers", "safetensors", "bart", "text2text-generation", "summarization", "en", "dataset:sengunsipahi/civitai_top10k", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
summarization
2024-03-20T00:27:44Z
--- language: en pipeline_tag: summarization tags: - summarization license: apache-2.0 metrics: - rouge model-index: - name: distilbart-cnn-12-6-finetuned-weaksup-1000 results: [] datasets: - sengunsipahi/civitai_top10k thumbnail: https://huggingface.co/front/thumbnails/distilbart_medium.png widget: - text: "pristine quality, White hair, detailed, bright green eyes, breezy, flowing hair, sunny, upper body, detailed face, summer, lush greenery, golden sunlight" context: "White hair, detailed bright green eyes, summer" --- ### Usage This checkpoint should be loaded into `BartForConditionalGeneration.from_pretrained`. See the [BART docs](https://huggingface.co/transformers/model_doc/bart.html?#transformers.BartForConditionalGeneration) for more information. # distilbart-cnn-12-6-SD-prompt This model is a [fine-tuned](https://pastebin.com/DTZ0WRz6) version of [sshleifer/distilbart-cnn-12-6](https://huggingface.co/sshleifer/distilbart-cnn-12-6) on an [dataset](https://huggingface.co/Nekochu/distilbart-cnn-12-6-SD-prompt/blob/main/dataset/dataset_CLIP.json), [modified](https://pastebin.com/6CVe3PMS) to be semi-synthetic by LLMs for summary Stable Diffusion Prompts.
dell-research-harvard/lt-wikidata-comp-fr
dell-research-harvard
2024-03-20T00:47:37Z
4
1
sentence-transformers
[ "sentence-transformers", "safetensors", "camembert", "linktransformer", "sentence-similarity", "tabular-classification", "fr", "arxiv:2309.00789", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
2023-08-10T20:37:20Z
--- pipeline_tag: sentence-similarity language: - fr tags: - linktransformer - sentence-transformers - sentence-similarity - tabular-classification --- # {MODEL_NAME} This is a [LinkTransformer](https://linktransformer.github.io/) model. At its core this model this is a sentence transformer model [sentence-transformers](https://www.SBERT.net) model- it just wraps around the class. It is designed for quick and easy record linkage (entity-matching) through the LinkTransformer package. The tasks include clustering, deduplication, linking, aggregation and more. Notwithstanding that, it can be used for any sentence similarity task within the sentence-transformers framework as well. It maps sentences & paragraphs to a 1024 dimensional dense vector space and can be used for tasks like clustering or semantic search. Take a look at the documentation of [sentence-transformers](https://www.sbert.net/index.html) if you want to use this model for more than what we support in our applications. This model has been fine-tuned on the model : dangvantuan/sentence-camembert-large. It is pretrained for the language : - fr. This model was trained on a dataset consisting of company aliases from wiki data using the LinkTransformer framework. It was trained for 100 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json ## Usage (LinkTransformer) Using this model becomes easy when you have [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) installed: ``` pip install -U linktransformer ``` Then you can use the model like this: ```python import linktransformer as lt import pandas as pd ##Load the two dataframes that you want to link. For example, 2 dataframes with company names that are written differently df1=pd.read_csv("data/df1.csv") ###This is the left dataframe with key CompanyName for instance df2=pd.read_csv("data/df2.csv") ###This is the right dataframe with key CompanyName for instance ###Merge the two dataframes on the key column! df_merged = lt.merge(df1, df2, on="CompanyName", how="inner") ##Done! The merged dataframe has a column called "score" that contains the similarity score between the two company names ``` ## Training your own LinkTransformer model Any Sentence Transformers can be used as a backbone by simply adding a pooling layer. Any other transformer on HuggingFace can also be used by specifying the option add_pooling_layer==True The model was trained using SupCon loss. Usage can be found in the package docs. The training config can be found in the repo with the name LT_training_config.json To replicate the training, you can download the file and specify the path in the config_path argument of the training function. You can also override the config by specifying the training_args argument. Here is an example. ```python ##Consider the example in the paper that has a dataset of Mexican products and their tariff codes from 1947 and 1948 and we want train a model to link the two tariff codes. saved_model_path = train_model( model_path="hiiamsid/sentence_similarity_spanish_es", dataset_path=dataset_path, left_col_names=["description47"], right_col_names=['description48'], left_id_name=['tariffcode47'], right_id_name=['tariffcode48'], log_wandb=False, config_path=LINKAGE_CONFIG_PATH, training_args={"num_epochs": 1} ) ``` You can also use this package for deduplication (clusters a df on the supplied key column). Merging a fine class (like product) to a coarse class (like HS code) is also possible. Read our paper and the documentation for more! ## Evaluation Results <!--- Describe how your model was evaluated --> You can evaluate the model using the [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) package's inference functions. We have provided a few datasets in the package for you to try out. We plan to host more datasets on Huggingface and our website (Coming soon) that you can take a look at. ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 664 with parameters: ``` {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `linktransformer.modified_sbert.losses.SupConLoss_wandb` Parameters of the fit()-Method: ``` { "epochs": 100, "evaluation_steps": 332, "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": null, "warmup_steps": 66400, "weight_decay": 0.01 } ``` LinkTransformer( (0): Transformer({'max_seq_length': 514, 'do_lower_case': False}) with Transformer model: CamembertModel (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False}) ) ``` ## Citing & Authors ``` @misc{arora2023linktransformer, title={LinkTransformer: A Unified Package for Record Linkage with Transformer Language Models}, author={Abhishek Arora and Melissa Dell}, year={2023}, eprint={2309.00789}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
Buseak/md_mt5_0109_v8
Buseak
2024-03-20T00:40:27Z
723
0
transformers
[ "transformers", "tensorboard", "safetensors", "mt5", "text2text-generation", "generated_from_trainer", "base_model:Buseak/md_mt5_0109_v7", "base_model:finetune:Buseak/md_mt5_0109_v7", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2024-03-19T21:12:01Z
--- license: apache-2.0 base_model: Buseak/md_mt5_0109_v7 tags: - generated_from_trainer metrics: - bleu model-index: - name: md_mt5_0109_v8 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # md_mt5_0109_v8 This model is a fine-tuned version of [Buseak/md_mt5_0109_v7](https://huggingface.co/Buseak/md_mt5_0109_v7) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0444 - Bleu: 0.6614 - Gen Len: 18.9444 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 15 ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:------:|:-------:| | 0.1129 | 1.0 | 975 | 0.0597 | 0.6517 | 18.9418 | | 0.1094 | 2.0 | 1950 | 0.0567 | 0.654 | 18.9372 | | 0.1101 | 3.0 | 2925 | 0.0543 | 0.657 | 18.9415 | | 0.1097 | 4.0 | 3900 | 0.0520 | 0.6555 | 18.9446 | | 0.1091 | 5.0 | 4875 | 0.0511 | 0.6571 | 18.9446 | | 0.1102 | 6.0 | 5850 | 0.0497 | 0.6591 | 18.9451 | | 0.1056 | 7.0 | 6825 | 0.0489 | 0.6585 | 18.9444 | | 0.1088 | 8.0 | 7800 | 0.0470 | 0.6595 | 18.9436 | | 0.1103 | 9.0 | 8775 | 0.0467 | 0.6589 | 18.9415 | | 0.1078 | 10.0 | 9750 | 0.0462 | 0.66 | 18.9423 | | 0.1106 | 11.0 | 10725 | 0.0451 | 0.6605 | 18.9431 | | 0.1112 | 12.0 | 11700 | 0.0448 | 0.6607 | 18.9444 | | 0.1134 | 13.0 | 12675 | 0.0447 | 0.6607 | 18.9395 | | 0.1183 | 14.0 | 13650 | 0.0446 | 0.6602 | 18.9408 | | 0.1188 | 15.0 | 14625 | 0.0444 | 0.6614 | 18.9444 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.2.1+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2
hyeogi/SOLAR-10.7B-v1.6
hyeogi
2024-03-20T00:38:31Z
2,243
1
transformers
[ "transformers", "safetensors", "llama", "text-generation", "SOLAR-10.7B", "conversational", "ko", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-20T00:23:36Z
--- language: - ko pipeline_tag: text-generation tags: - SOLAR-10.7B license: cc-by-nc-4.0 --- # SOLAR-10.7B ### Model Details - Base Model: [yanolja/KoSOLAR-10.7B-v0.2](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.2) ### Datasets - sampling and translate [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca) - sampling and instrcution format [HAERAE-HUB/KMMLU](https://huggingface.co/datasets/HAERAE-HUB/KMMLU)
mvpmaster/nddmp-kellemar-KrishnaHercules-7b-slerp
mvpmaster
2024-03-20T00:36:31Z
4
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "mvpmaster/NeuralDareDMistralPro-7b-slerp", "mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp", "base_model:mvpmaster/NeuralDareDMistralPro-7b-slerp", "base_model:merge:mvpmaster/NeuralDareDMistralPro-7b-slerp", "base_model:mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp", "base_model:merge:mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-20T00:32:26Z
--- tags: - merge - mergekit - lazymergekit - mvpmaster/NeuralDareDMistralPro-7b-slerp - mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp base_model: - mvpmaster/NeuralDareDMistralPro-7b-slerp - mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp --- # nddmp-kellemar-KrishnaHercules-7b-slerp nddmp-kellemar-KrishnaHercules-7b-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [mvpmaster/NeuralDareDMistralPro-7b-slerp](https://huggingface.co/mvpmaster/NeuralDareDMistralPro-7b-slerp) * [mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp](https://huggingface.co/mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp) ## 🧩 Configuration ```yaml slices: - sources: - model: mvpmaster/NeuralDareDMistralPro-7b-slerp layer_range: [0, 32] - model: mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp layer_range: [0, 32] merge_method: slerp base_model: mvpmaster/NeuralDareDMistralPro-7b-slerp parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "mvpmaster/nddmp-kellemar-KrishnaHercules-7b-slerp" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
Denath-Khor/ARIA-7B-V3-mistral-french-v1-GGUF
Denath-Khor
2024-03-20T00:28:38Z
2
0
null
[ "gguf", "fr", "en", "dataset:open-llm-leaderboard/details_Faradaylab__ARIA-70B-V3", "endpoints_compatible", "region:us", "conversational" ]
null
2024-03-19T22:21:45Z
--- datasets: - open-llm-leaderboard/details_Faradaylab__ARIA-70B-V3 language: - fr - en --- Fenris MOBILE GGUF - Q4_K_M - Q8_0 ARIA-7B-V3-mistral-french-v1 - GGUF Model creator : Faradaylab Original model : ARIA-7B-V3-mistral-french-v1 : https://huggingface.co/Faradaylab/ARIA-7B-V3-mistral-french-v1 Description : This repo contains GGUF format model files for Faradaylab's ARIA-7B-V3-mistral-french-v1. Finetuned from : mistralai/Mistral-7B-v0.1
dell-research-harvard/lt-wikidata-comp-en
dell-research-harvard
2024-03-20T00:27:30Z
18,812
2
sentence-transformers
[ "sentence-transformers", "safetensors", "mpnet", "linktransformer", "sentence-similarity", "tabular-classification", "en", "arxiv:2309.00789", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
2023-08-11T04:08:15Z
--- pipeline_tag: sentence-similarity language: - en tags: - linktransformer - sentence-transformers - sentence-similarity - tabular-classification --- # {MODEL_NAME} This is a [LinkTransformer](https://linktransformer.github.io/) model. At its core this model this is a sentence transformer model [sentence-transformers](https://www.SBERT.net) model- it just wraps around the class. It is designed for quick and easy record linkage (entity-matching) through the LinkTransformer package. The tasks include clustering, deduplication, linking, aggregation and more. Notwithstanding that, it can be used for any sentence similarity task within the sentence-transformers framework as well. It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. Take a look at the documentation of [sentence-transformers](https://www.sbert.net/index.html) if you want to use this model for more than what we support in our applications. This model has been fine-tuned on the model : multi-qa-mpnet-base-dot-v1. It is pretrained for the language : - en. This model was trained on a dataset consisting of company aliases from wiki data using the LinkTransformer framework. It was trained for 100 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json ## Usage (LinkTransformer) Using this model becomes easy when you have [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) installed: ``` pip install -U linktransformer ``` Then you can use the model like this: ```python import linktransformer as lt import pandas as pd ##Load the two dataframes that you want to link. For example, 2 dataframes with company names that are written differently df1=pd.read_csv("data/df1.csv") ###This is the left dataframe with key CompanyName for instance df2=pd.read_csv("data/df2.csv") ###This is the right dataframe with key CompanyName for instance ###Merge the two dataframes on the key column! df_merged = lt.merge(df1, df2, on="CompanyName", how="inner") ##Done! The merged dataframe has a column called "score" that contains the similarity score between the two company names ``` ## Training your own LinkTransformer model Any Sentence Transformers can be used as a backbone by simply adding a pooling layer. Any other transformer on HuggingFace can also be used by specifying the option add_pooling_layer==True The model was trained using SupCon loss. Usage can be found in the package docs. The training config can be found in the repo with the name LT_training_config.json To replicate the training, you can download the file and specify the path in the config_path argument of the training function. You can also override the config by specifying the training_args argument. Here is an example. ```python ##Consider the example in the paper that has a dataset of Mexican products and their tariff codes from 1947 and 1948 and we want train a model to link the two tariff codes. saved_model_path = train_model( model_path="hiiamsid/sentence_similarity_spanish_es", dataset_path=dataset_path, left_col_names=["description47"], right_col_names=['description48'], left_id_name=['tariffcode47'], right_id_name=['tariffcode48'], log_wandb=False, config_path=LINKAGE_CONFIG_PATH, training_args={"num_epochs": 1} ) ``` You can also use this package for deduplication (clusters a df on the supplied key column). Merging a fine class (like product) to a coarse class (like HS code) is also possible. Read our paper and the documentation for more! ## Evaluation Results <!--- Describe how your model was evaluated --> You can evaluate the model using the [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) package's inference functions. We have provided a few datasets in the package for you to try out. We plan to host more datasets on Huggingface and our website (Coming soon) that you can take a look at. ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 2087 with parameters: ``` {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `linktransformer.modified_sbert.losses.SupConLoss_wandb` Parameters of the fit()-Method: ``` { "epochs": 100, "evaluation_steps": 1044, "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": null, "warmup_steps": 208700, "weight_decay": 0.01 } ``` LinkTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: MPNetModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False}) ) ``` ## Citing & Authors ``` @misc{arora2023linktransformer, title={LinkTransformer: A Unified Package for Record Linkage with Transformer Language Models}, author={Abhishek Arora and Melissa Dell}, year={2023}, eprint={2309.00789}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
dell-research-harvard/lt-wikidata-comp-prod-ind-ja
dell-research-harvard
2024-03-20T00:21:49Z
6
1
sentence-transformers
[ "sentence-transformers", "safetensors", "luke", "linktransformer", "sentence-similarity", "tabular-classification", "ja", "arxiv:2309.00789", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
2024-01-30T06:29:34Z
--- pipeline_tag: sentence-similarity language: - ja tags: - linktransformer - sentence-transformers - sentence-similarity - tabular-classification --- # {MODEL_NAME} This is a [LinkTransformer](https://linktransformer.github.io/) model. At its core this model this is a sentence transformer model [sentence-transformers](https://www.SBERT.net) model- it just wraps around the class. It is designed for quick and easy record linkage (entity-matching) through the LinkTransformer package. The tasks include clustering, deduplication, linking, aggregation and more. Notwithstanding that, it can be used for any sentence similarity task within the sentence-transformers framework as well. It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. Take a look at the documentation of [sentence-transformers](https://www.sbert.net/index.html) if you want to use this model for more than what we support in our applications. This model has been fine-tuned on the model : oshizo/sbert-jsnli-luke-japanese-base-lite. It is pretrained for the language : - ja. This is a (Modern) Japanese Link Transformer model - trained on Company <SEP> Product <SEP> Industry from wiki data. ## Usage (LinkTransformer) Using this model becomes easy when you have [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) installed: ``` pip install -U linktransformer ``` Then you can use the model like this: ```python import linktransformer as lt import pandas as pd ##Load the two dataframes that you want to link. For example, 2 dataframes with company names that are written differently df1=pd.read_csv("data/df1.csv") ###This is the left dataframe with key CompanyName for instance df2=pd.read_csv("data/df2.csv") ###This is the right dataframe with key CompanyName for instance ###Merge the two dataframes on the key column! df_merged = lt.merge(df1, df2, on="CompanyName", how="inner") ##Done! The merged dataframe has a column called "score" that contains the similarity score between the two company names ``` ## Training your own LinkTransformer model Any Sentence Transformers can be used as a backbone by simply adding a pooling layer. Any other transformer on HuggingFace can also be used by specifying the option add_pooling_layer==True The model was trained using SupCon loss. Usage can be found in the package docs. The training config can be found in the repo with the name LT_training_config.json To replicate the training, you can download the file and specify the path in the config_path argument of the training function. You can also override the config by specifying the training_args argument. Here is an example. ```python ##Consider the example in the paper that has a dataset of Mexican products and their tariff codes from 1947 and 1948 and we want train a model to link the two tariff codes. saved_model_path = train_model( model_path="hiiamsid/sentence_similarity_spanish_es", dataset_path=dataset_path, left_col_names=["description47"], right_col_names=['description48'], left_id_name=['tariffcode47'], right_id_name=['tariffcode48'], log_wandb=False, config_path=LINKAGE_CONFIG_PATH, training_args={"num_epochs": 1} ) ``` You can also use this package for deduplication (clusters a df on the supplied key column). Merging a fine class (like product) to a coarse class (like HS code) is also possible. Read our paper and the documentation for more! ## Evaluation Results <!--- Describe how your model was evaluated --> You can evaluate the model using the [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) package's inference functions. We have provided a few datasets in the package for you to try out. We plan to host more datasets on Huggingface and our website (Coming soon) that you can take a look at. ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 57 with parameters: ``` {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `linktransformer.modified_sbert.losses.SupConLoss_wandb` Parameters of the fit()-Method: ``` { "epochs": 70, "evaluation_steps": 29, "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": null, "warmup_steps": 3990, "weight_decay": 0.01 } ``` LinkTransformer( (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: LukeModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False}) ) ``` ## Citing & Authors ``` @misc{arora2023linktransformer, title={LinkTransformer: A Unified Package for Record Linkage with Transformer Language Models}, author={Abhishek Arora and Melissa Dell}, year={2023}, eprint={2309.00789}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
dell-research-harvard/lt-historicjapan-onlinecontrastive
dell-research-harvard
2024-03-20T00:19:22Z
4
1
sentence-transformers
[ "sentence-transformers", "safetensors", "luke", "linktransformer", "sentence-similarity", "tabular-classification", "ja", "arxiv:2309.00789", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
2024-03-20T00:19:15Z
--- pipeline_tag: sentence-similarity language: - ja tags: - linktransformer - sentence-transformers - sentence-similarity - tabular-classification --- # {MODEL_NAME} This is a [LinkTransformer](https://linktransformer.github.io/) model. At its core this model this is a sentence transformer model [sentence-transformers](https://www.SBERT.net) model- it just wraps around the class. It is designed for quick and easy record linkage (entity-matching) through the LinkTransformer package. The tasks include clustering, deduplication, linking, aggregation and more. Notwithstanding that, it can be used for any sentence similarity task within the sentence-transformers framework as well. It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. Take a look at the documentation of [sentence-transformers](https://www.sbert.net/index.html) if you want to use this model for more than what we support in our applications. This model has been fine-tuned on the model : oshizo/sbert-jsnli-luke-japanese-base-lite. It is pretrained for the language : - ja. This model was trained on a dataset of historic Japanese companies, products, industry, addresses, and shareholders. Take a look at our paper for more details. The task is to link indices of japanese companies ## Usage (LinkTransformer) Using this model becomes easy when you have [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) installed: ``` pip install -U linktransformer ``` Then you can use the model like this: ```python import linktransformer as lt import pandas as pd ##Load the two dataframes that you want to link. For example, 2 dataframes with company names that are written differently df1=pd.read_csv("data/df1.csv") ###This is the left dataframe with key CompanyName for instance df2=pd.read_csv("data/df2.csv") ###This is the right dataframe with key CompanyName for instance ###Merge the two dataframes on the key column! df_merged = lt.merge(df1, df2, on="CompanyName", how="inner") ##Done! The merged dataframe has a column called "score" that contains the similarity score between the two company names ``` ## Training your own LinkTransformer model Any Sentence Transformers can be used as a backbone by simply adding a pooling layer. Any other transformer on HuggingFace can also be used by specifying the option add_pooling_layer==True The model was trained using SupCon loss. Usage can be found in the package docs. The training config can be found in the repo with the name LT_training_config.json To replicate the training, you can download the file and specify the path in the config_path argument of the training function. You can also override the config by specifying the training_args argument. Here is an example. ```python ##Consider the example in the paper that has a dataset of Mexican products and their tariff codes from 1947 and 1948 and we want train a model to link the two tariff codes. saved_model_path = train_model( model_path="hiiamsid/sentence_similarity_spanish_es", dataset_path=dataset_path, left_col_names=["description47"], right_col_names=['description48'], left_id_name=['tariffcode47'], right_id_name=['tariffcode48'], log_wandb=False, config_path=LINKAGE_CONFIG_PATH, training_args={"num_epochs": 1} ) ``` You can also use this package for deduplication (clusters a df on the supplied key column). Merging a fine class (like product) to a coarse class (like HS code) is also possible. Read our paper and the documentation for more! ## Evaluation Results <!--- Describe how your model was evaluated --> You can evaluate the model using the [LinkTransformer](https://github.com/dell-research-harvard/linktransformer) package's inference functions. We have provided a few datasets in the package for you to try out. We plan to host more datasets on Huggingface and our website (Coming soon) that you can take a look at. ## Training The model was trained with the parameters: **DataLoader**: `torch.utils.data.dataloader.DataLoader` of length 45 with parameters: ``` {'batch_size': 64, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'} ``` **Loss**: `linktransformer.modified_sbert.losses.OnlineContrastiveLoss_wandb` Parameters of the fit()-Method: ``` { "epochs": 50, "evaluation_steps": 23, "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator", "max_grad_norm": 1, "optimizer_class": "<class 'torch.optim.adamw.AdamW'>", "optimizer_params": { "lr": 2e-05 }, "scheduler": "WarmupLinear", "steps_per_epoch": null, "warmup_steps": 2250, "weight_decay": 0.01 } ``` LinkTransformer( (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: LukeModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False}) ) ``` ## Citing & Authors ``` @misc{arora2023linktransformer, title={LinkTransformer: A Unified Package for Record Linkage with Transformer Language Models}, author={Abhishek Arora and Melissa Dell}, year={2023}, eprint={2309.00789}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```
NGDJ/Mistral-AI-PPA-16bit
NGDJ
2024-03-20T00:16:51Z
39
0
transformers
[ "transformers", "gguf", "mistral", "text-generation-inference", "unsloth", "en", "base_model:unsloth/mistral-7b-bnb-4bit", "base_model:quantized:unsloth/mistral-7b-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-03-20T00:05:56Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - mistral - gguf base_model: unsloth/mistral-7b-bnb-4bit --- # Uploaded model - **Developed by:** NGDJ - **License:** apache-2.0 - **Finetuned from model :** unsloth/mistral-7b-bnb-4bit This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
rishikesanr/stack-overflow-bot-llama2
rishikesanr
2024-03-20T00:13:58Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-03-20T00:13:37Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
linoyts/huggy_lora_v2_unet
linoyts
2024-03-20T00:07:17Z
2
0
diffusers
[ "diffusers", "stable-diffusion-xl", "stable-diffusion-xl-diffusers", "diffusers-training", "text-to-image", "lora", "template:sd-lora", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "base_model:adapter:stabilityai/stable-diffusion-xl-base-1.0", "license:openrail++", "region:us" ]
text-to-image
2024-03-19T23:30:11Z
--- tags: - stable-diffusion-xl - stable-diffusion-xl-diffusers - diffusers-training - text-to-image - diffusers - lora - template:sd-lora widget: - text: 'a TOK emoji dressed as yoda' output: url: "image_0.png" - text: 'a TOK emoji dressed as yoda' output: url: "image_1.png" - text: 'a TOK emoji dressed as yoda' output: url: "image_2.png" - text: 'a TOK emoji dressed as yoda' output: url: "image_3.png" base_model: stabilityai/stable-diffusion-xl-base-1.0 instance_prompt: a TOK emoji license: openrail++ --- # SDXL LoRA DreamBooth - linoyts/huggy_lora_v2_unet <Gallery /> ## Model description ### These are linoyts/huggy_lora_v2_unet LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0. ## Download model ### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke - **LoRA**: download **[`huggy_lora_v2_unet.safetensors` here 💾](/linoyts/huggy_lora_v2_unet/blob/main/huggy_lora_v2_unet.safetensors)**. - Place it on your `models/Lora` folder. - On AUTOMATIC1111, load the LoRA by adding `<lora:huggy_lora_v2_unet:1>` to your prompt. On ComfyUI just [load it as a regular LoRA](https://comfyanonymous.github.io/ComfyUI_examples/lora/). ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('stabilityai/stable-diffusion-xl-base-1.0', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('linoyts/huggy_lora_v2_unet', weight_name='pytorch_lora_weights.safetensors') image = pipeline('a TOK emoji dressed as yoda').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters) ## Trigger words You should use a TOK emoji to trigger the image generation. ## Details All [Files & versions](/linoyts/huggy_lora_v2_unet/tree/main). The weights were trained using [🧨 diffusers Advanced Dreambooth Training Script](https://github.com/huggingface/diffusers/blob/main/examples/advanced_diffusion_training/train_dreambooth_lora_sdxl_advanced.py). LoRA for the text encoder was enabled. False. Pivotal tuning was enabled: False. Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
jeiku/parttwo
jeiku
2024-03-20T00:06:46Z
6
1
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T23:56:10Z
--- base_model: [] library_name: transformers tags: - mergekit - merge --- # parttwo This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the SLERP merge method. ### Models Merged The following models were included in the merge: * first * second ### Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - model: first layer_range: [0, 32] - model: second layer_range: [0, 32] merge_method: slerp base_model: first parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ```
NGDJ/Mistral-AI-PPA
NGDJ
2024-03-20T00:01:53Z
4
0
transformers
[ "transformers", "gguf", "mistral", "text-generation-inference", "unsloth", "en", "base_model:unsloth/mistral-7b-bnb-4bit", "base_model:quantized:unsloth/mistral-7b-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-03-19T23:58:27Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - mistral - gguf base_model: unsloth/mistral-7b-bnb-4bit --- # Uploaded model - **Developed by:** NGDJ - **License:** apache-2.0 - **Finetuned from model :** unsloth/mistral-7b-bnb-4bit This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
mvpmaster/NeuralDareDMistralPro-7b-slerp
mvpmaster
2024-03-19T23:59:52Z
58
1
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "mlabonne/NeuralDaredevil-7B", "NousResearch/Hermes-2-Pro-Mistral-7B", "base_model:NousResearch/Hermes-2-Pro-Mistral-7B", "base_model:merge:NousResearch/Hermes-2-Pro-Mistral-7B", "base_model:mlabonne/NeuralDaredevil-7B", "base_model:merge:mlabonne/NeuralDaredevil-7B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-18T23:52:49Z
--- license: apache-2.0 tags: - merge - mergekit - lazymergekit - mlabonne/NeuralDaredevil-7B - NousResearch/Hermes-2-Pro-Mistral-7B base_model: - mlabonne/NeuralDaredevil-7B - NousResearch/Hermes-2-Pro-Mistral-7B --- # NeuralDareDMistralPro-slerp NeuralDareDMistralPro-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [mlabonne/NeuralDaredevil-7B](https://huggingface.co/mlabonne/NeuralDaredevil-7B) * [NousResearch/Hermes-2-Pro-Mistral-7B](https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B) ## 🧩 Configuration ```yaml slices: - sources: - model: mlabonne/NeuralDaredevil-7B layer_range: [0, 32] - model: NousResearch/Hermes-2-Pro-Mistral-7B layer_range: [0, 32] merge_method: slerp base_model: mlabonne/NeuralDaredevil-7B parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "mvpmaster/NeuralDareDMistralPro-slerp" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
jeiku/partone
jeiku
2024-03-19T23:54:15Z
4
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T23:35:28Z
--- base_model: [] library_name: transformers tags: - mergekit - merge --- # one This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the SLERP merge method. ### Models Merged The following models were included in the merge: * one * two ### Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - model: one layer_range: [0, 32] - model: two layer_range: [0, 32] merge_method: slerp base_model: two parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ```
ldowey/llama-2-7b-sentinomics_2
ldowey
2024-03-19T23:49:36Z
4
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T23:42:49Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
alterf/det_v2
alterf
2024-03-19T23:46:18Z
188
0
transformers
[ "transformers", "safetensors", "detr", "image-feature-extraction", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
image-feature-extraction
2024-03-19T23:23:56Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
NGDJ/Mistral-7B-Summarization-QLoRa
NGDJ
2024-03-19T23:42:55Z
4
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "text-generation-inference", "unsloth", "trl", "en", "base_model:unsloth/mistral-7b-bnb-4bit", "base_model:finetune:unsloth/mistral-7b-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T23:37:34Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - mistral - trl base_model: unsloth/mistral-7b-bnb-4bit --- # Uploaded model - **Developed by:** NGDJ - **License:** apache-2.0 - **Finetuned from model :** unsloth/mistral-7b-bnb-4bit This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
minhah/vivit-b-16x2-kinetics400-finetuned-elder
minhah
2024-03-19T23:39:24Z
65
0
transformers
[ "transformers", "safetensors", "vivit", "video-classification", "generated_from_trainer", "base_model:google/vivit-b-16x2-kinetics400", "base_model:finetune:google/vivit-b-16x2-kinetics400", "license:mit", "endpoints_compatible", "region:us" ]
video-classification
2024-03-19T14:27:01Z
--- license: mit base_model: google/vivit-b-16x2-kinetics400 tags: - generated_from_trainer metrics: - accuracy model-index: - name: vivit-b-16x2-kinetics400-finetuned-elder results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vivit-b-16x2-kinetics400-finetuned-elder This model is a fine-tuned version of [google/vivit-b-16x2-kinetics400](https://huggingface.co/google/vivit-b-16x2-kinetics400) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.6807 - Accuracy: 0.3205 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - training_steps: 576 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.6183 | 0.25 | 145 | 1.6139 | 0.3360 | | 1.5777 | 1.25 | 290 | 1.6061 | 0.3024 | | 1.36 | 2.25 | 435 | 1.6442 | 0.2863 | | 1.5395 | 3.24 | 576 | 1.6518 | 0.2688 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.1.0+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2
jodchen/llm_lora
jodchen
2024-03-19T23:32:14Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-03-19T21:07:03Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
linoyts/huggy_dora_v3_unet
linoyts
2024-03-19T23:29:52Z
3
0
diffusers
[ "diffusers", "stable-diffusion-xl", "stable-diffusion-xl-diffusers", "diffusers-training", "text-to-image", "dora", "template:sd-lora", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "base_model:finetune:stabilityai/stable-diffusion-xl-base-1.0", "license:openrail++", "region:us" ]
text-to-image
2024-03-19T22:39:44Z
--- tags: - stable-diffusion-xl - stable-diffusion-xl-diffusers - diffusers-training - text-to-image - diffusers - dora - template:sd-lora widget: - text: 'a TOK emoji dressed as yoda' output: url: "image_0.png" - text: 'a TOK emoji dressed as yoda' output: url: "image_1.png" - text: 'a TOK emoji dressed as yoda' output: url: "image_2.png" - text: 'a TOK emoji dressed as yoda' output: url: "image_3.png" base_model: stabilityai/stable-diffusion-xl-base-1.0 instance_prompt: a TOK emoji license: openrail++ --- # SDXL LoRA DreamBooth - linoyts/huggy_dora_v3_unet <Gallery /> ## Model description ### These are linoyts/huggy_dora_v3_unet LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0. ## Download model ### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke - **LoRA**: download **[`huggy_dora_v3_unet.safetensors` here 💾](/linoyts/huggy_dora_v3_unet/blob/main/huggy_dora_v3_unet.safetensors)**. - Place it on your `models/Lora` folder. - On AUTOMATIC1111, load the LoRA by adding `<lora:huggy_dora_v3_unet:1>` to your prompt. On ComfyUI just [load it as a regular LoRA](https://comfyanonymous.github.io/ComfyUI_examples/lora/). ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('stabilityai/stable-diffusion-xl-base-1.0', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('linoyts/huggy_dora_v3_unet', weight_name='pytorch_lora_weights.safetensors') image = pipeline('a TOK emoji dressed as yoda').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters) ## Trigger words You should use a TOK emoji to trigger the image generation. ## Details All [Files & versions](/linoyts/huggy_dora_v3_unet/tree/main). The weights were trained using [🧨 diffusers Advanced Dreambooth Training Script](https://github.com/huggingface/diffusers/blob/main/examples/advanced_diffusion_training/train_dreambooth_lora_sdxl_advanced.py). LoRA for the text encoder was enabled. False. Pivotal tuning was enabled: False. Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
alterf/detp_v2
alterf
2024-03-19T23:24:04Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-03-19T23:24:04Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
mvpmaster/Einstein-4D-Marcoro14-12b-32k-experiment
mvpmaster
2024-03-19T23:18:12Z
6
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp", "base_model:mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp", "base_model:finetune:mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T23:11:19Z
--- tags: - merge - mergekit - lazymergekit - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp base_model: - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp --- # Einstein-4D-Marcoro14-12b-32k-experiment Einstein-4D-Marcoro14-12b-32k-experiment is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp](https://huggingface.co/mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp) * [mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp](https://huggingface.co/mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp) * [mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp](https://huggingface.co/mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp) * [mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp](https://huggingface.co/mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp) * [mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp](https://huggingface.co/mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp) * [mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp](https://huggingface.co/mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp) * [mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp](https://huggingface.co/mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp) ## 🧩 Configuration ```yaml dtype: float16 merge_method: passthrough slices: - sources: - layer_range: [0, 8] model: mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - sources: - layer_range: [4, 12] model: mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - sources: - layer_range: [8, 16] model: mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - sources: - layer_range: [12, 20] model: mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - sources: - layer_range: [16, 24] model: mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - sources: - layer_range: [20, 28] model: mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp - sources: - layer_range: [24, 32] model: mvpmaster/Einstein-4D-Marcoro14-7b-full-slerp ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "mvpmaster/Einstein-4D-Marcoro14-12b-32k-experiment" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
ND911/Franken-MistressMaid-10.5B-v2
ND911
2024-03-19T23:09:37Z
6
1
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T22:12:26Z
--- base_model: [] library_name: transformers tags: - mergekit - merge --- ![](mistressmaid.png) # Franken-Mistress-10.5B-v2 This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details So far of the Franken merges, this one does very well using the Min-P and Noromaid settings in SillyTavern 2. This one seems even better then the 10.5B version of this model. I uploaded 3 files for SillyTavern that can be imported. I take no credit for these files, not sure who original authors are. * [MinP-text-completion-preset.json](https://huggingface.co/ND911/Franken-MistressMaid-7B-v2/blob/main/MinP-text-completion-preset.json]) * [rp-merge-text-completion-preset.json](https://huggingface.co/ND911/Franken-MistressMaid-7B-v2/blob/main/rp-merge-text-completion-preset.json) * [noromaid-context-template.json](https://huggingface.co/ND911/Franken-MistressMaid-7B-v2/blob/main/noromaid-context-template.json) ### Merge Method This model was merged using the passthrough merge method. ### Models Merged The following models were included in the merge: * Franken-Maid-v2 ### Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - model: Franken-Maid-v2 layer_range: [0, 16] - sources: - model: Franken-Maid-v2 layer_range: [8, 24] - sources: - model: Franken-Maid-v2 layer_range: [17, 32] merge_method: passthrough dtype: float16 ``` ```yaml models: - model: ibm/merlinite-7b parameters: weight: 1 density: 1 - model: Undi95/Toppy-M-7B parameters: weight: 0.3 - model: jondurbin/bagel-dpo-7b-v0.4 parameters: weight: 0.2 - model: senseable/WestLake-7B-v2 parameters: weight: 0.2 - model: l3utterfly/mistral-7b-v0.1-layla-v4 parameters: weight: 0.2 merge_method: ties base_model: Franken-Maid parameters: density: 0.4 int8_mask: true normalize: true dtype: bfloat16 ``` ```yaml models: - model: SanjiWatsuki/Sonya-7B parameters: weight: 1 density: 1 - model: SanjiWatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE parameters: weight: 0.3 - model: Azazelle/Half-NSFW_Noromaid-7b parameters: weight: 0.2 - model: senseable/WestLake-7B-v2 parameters: weight: 0.2 - model: l3utterfly/mistral-7b-v0.1-layla-v4 parameters: weight: 0.2 merge_method: ties base_model: Weyaxi/OpenHermes-2.5-neural-chat-7b-v3-1-7B parameters: density: 0.4 int8_mask: true normalize: true dtype: bfloat16 ```
sarthakharne/bert-base-125-ep-pretrain-on-textbooks
sarthakharne
2024-03-19T23:03:47Z
179
0
transformers
[ "transformers", "safetensors", "bert", "fill-mask", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2024-03-19T23:02:12Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Kryptone/GPTSVC
Kryptone
2024-03-19T22:48:09Z
0
0
null
[ "license:openrail++", "region:us" ]
null
2024-03-19T22:35:23Z
--- license: openrail++ --- # GPTSVC (GPT So-Vits Collection) <!-- Provide a quick summary of what the model is/does. --> This is a collection of all my models trained using GPT So-Vits. All models in here will mostly be Japanese unless otherwise noted. ## No models in here will work with w-okada, an official one is in development by RVC-Boss (maybe), so when/if it does come out, this description will be updated.
Digoguima/Djavanmodel
Digoguima
2024-03-19T22:45:33Z
0
0
diffusers
[ "diffusers", "text-to-image", "stable-diffusion", "lora", "template:sd-lora", "base_model:ByteDance/SDXL-Lightning", "base_model:adapter:ByteDance/SDXL-Lightning", "region:us" ]
text-to-image
2024-03-19T22:45:11Z
--- tags: - text-to-image - stable-diffusion - lora - diffusers - template:sd-lora widget: - text: '-' output: url: images/1000388922.jpg base_model: ByteDance/SDXL-Lightning instance_prompt: null --- # Djavan <Gallery /> ## Download model [Download](/Digoguima/Djavan/tree/main) them in the Files & versions tab.
XiShi5941/llama-2-7b-pdtb2.0-epoch3-p4-fix20240319161911
XiShi5941
2024-03-19T22:45:12Z
4
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T22:42:17Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
linoyts/huggy_lora_v3_unet
linoyts
2024-03-19T22:39:26Z
4
0
diffusers
[ "diffusers", "stable-diffusion-xl", "stable-diffusion-xl-diffusers", "diffusers-training", "text-to-image", "lora", "template:sd-lora", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "base_model:adapter:stabilityai/stable-diffusion-xl-base-1.0", "license:openrail++", "region:us" ]
text-to-image
2024-03-19T22:04:49Z
--- tags: - stable-diffusion-xl - stable-diffusion-xl-diffusers - diffusers-training - text-to-image - diffusers - lora - template:sd-lora widget: - text: 'a TOK emoji dressed as yoda' output: url: "image_0.png" - text: 'a TOK emoji dressed as yoda' output: url: "image_1.png" - text: 'a TOK emoji dressed as yoda' output: url: "image_2.png" - text: 'a TOK emoji dressed as yoda' output: url: "image_3.png" base_model: stabilityai/stable-diffusion-xl-base-1.0 instance_prompt: a TOK emoji license: openrail++ --- # SDXL LoRA DreamBooth - linoyts/huggy_lora_v3_unet <Gallery /> ## Model description ### These are linoyts/huggy_lora_v3_unet LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0. ## Download model ### Use it with UIs such as AUTOMATIC1111, Comfy UI, SD.Next, Invoke - **LoRA**: download **[`huggy_lora_v3_unet.safetensors` here 💾](/linoyts/huggy_lora_v3_unet/blob/main/huggy_lora_v3_unet.safetensors)**. - Place it on your `models/Lora` folder. - On AUTOMATIC1111, load the LoRA by adding `<lora:huggy_lora_v3_unet:1>` to your prompt. On ComfyUI just [load it as a regular LoRA](https://comfyanonymous.github.io/ComfyUI_examples/lora/). ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('stabilityai/stable-diffusion-xl-base-1.0', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('linoyts/huggy_lora_v3_unet', weight_name='pytorch_lora_weights.safetensors') image = pipeline('a TOK emoji dressed as yoda').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters) ## Trigger words You should use a TOK emoji to trigger the image generation. ## Details All [Files & versions](/linoyts/huggy_lora_v3_unet/tree/main). The weights were trained using [🧨 diffusers Advanced Dreambooth Training Script](https://github.com/huggingface/diffusers/blob/main/examples/advanced_diffusion_training/train_dreambooth_lora_sdxl_advanced.py). LoRA for the text encoder was enabled. False. Pivotal tuning was enabled: False. Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
bisoye/distilbert-base-uncased-finetuned-clinc
bisoye
2024-03-19T22:38:07Z
5
0
transformers
[ "transformers", "tensorboard", "safetensors", "distilbert", "text-classification", "generated_from_trainer", "dataset:clinc_oos", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-03-19T19:58:50Z
--- license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer datasets: - clinc_oos metrics: - accuracy model-index: - name: distilbert-base-uncased-finetuned-clinc results: - task: name: Text Classification type: text-classification dataset: name: clinc_oos type: clinc_oos config: plus split: validation args: plus metrics: - name: Accuracy type: accuracy value: 0.9135483870967742 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-clinc This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the clinc_oos dataset. It achieves the following results on the evaluation set: - Loss: 0.8068 - Accuracy: 0.9135 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 48 - eval_batch_size: 48 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 318 | 3.3144 | 0.7206 | | 3.8129 | 2.0 | 636 | 1.9134 | 0.8474 | | 3.8129 | 3.0 | 954 | 1.1920 | 0.8855 | | 1.7365 | 4.0 | 1272 | 0.8920 | 0.9113 | | 0.9362 | 5.0 | 1590 | 0.8068 | 0.9135 | ### Framework versions - Transformers 4.38.1 - Pytorch 2.1.2 - Datasets 2.1.0 - Tokenizers 0.15.2
yehiawp4/vivit-b-16x2-kinetics400-finetuned-caer-subset
yehiawp4
2024-03-19T22:29:32Z
65
0
transformers
[ "transformers", "safetensors", "vivit", "video-classification", "generated_from_trainer", "base_model:google/vivit-b-16x2-kinetics400", "base_model:finetune:google/vivit-b-16x2-kinetics400", "license:mit", "endpoints_compatible", "region:us" ]
video-classification
2024-03-19T21:33:30Z
--- license: mit base_model: google/vivit-b-16x2-kinetics400 tags: - generated_from_trainer metrics: - accuracy model-index: - name: vivit-b-16x2-kinetics400-finetuned-caer-subset results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vivit-b-16x2-kinetics400-finetuned-caer-subset This model is a fine-tuned version of [google/vivit-b-16x2-kinetics400](https://huggingface.co/google/vivit-b-16x2-kinetics400) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.8413 - Accuracy: 0.2330 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - training_steps: 350 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 1.8092 | 0.56 | 196 | 1.9746 | 0.1707 | | 1.4353 | 1.44 | 350 | 1.8480 | 0.2439 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.1.0 - Datasets 2.18.0 - Tokenizers 0.15.2
adjohn1313/wizard_sft_explainable_rlhf_6k
adjohn1313
2024-03-19T22:23:20Z
74
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "awq", "region:us" ]
text-generation
2024-03-17T20:02:23Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
SWBFSpy/SWBF_Voiceovers_Yoda_Palpatine_Dooku_Ackbar
SWBFSpy
2024-03-19T22:22:58Z
0
0
null
[ "region:us" ]
null
2024-03-19T21:46:15Z
## SWBF Campaign Voiceover Generators (Yoda, Palpatine, Dooku, Ackbar) ![image/jpeg](https://i.imgur.com/vW3Qfk0.jpeg) RVC neural network models of the four voice actors from the original Star Wars: Battlefront (2004) game -- Master Yoda, Senator Palpatine, Count Dooku and Admiral Ackbar. Trained on 1000 epochs. This gives us a pool of "virtual voice actors" to generate mission briefing shell/core scripts for all the hundreds of new maps our mod team is adding to the SWBF campaign (online and singleplayer compatible). In a way, this tool helps us resurrect the dev team of original voice actors, so we can finish building many of the ideas that were abandoned from the original 6/13/2003 design document for Star Wars: The Front Line, in addition to other new features. We will expand this pack later with Luke, Vader, Windu, commander and team voices, etc. This ZeroBuilder tool created by Phobos allows you to add your own voiceover extensions to the stock campaigns. You should give credit to SWBFmodders, Phobos, LucasArts, Pandemic Studios LLC and the original SWBF voice actors if you use these for your mods. ### SWBF Voice Actors G.W. Childs - Imperial Infantry<br> Chris Cox - Alliance Infantry, CIS Infantry, CIS Officer, Gungan Infantry<br> Nick Jamison - Alliance Officer, Darth Sidious, Emperor Palpatine<br> Tom Kane - Admiral Ackbar, Yoda<br> Temuera Morrison - Republic Infantry, Republic Officer<br> David Robb - Imperial Officer<br> ### Notice These exist as proof of concept. There are alternate VOs available such as<br> https://huggingface.co/bowlql/YodaRVC<br> https://huggingface.co/mthxz/palpatine<br> https://huggingface.co/Akitai/DookuCGI<br> https://huggingface.co/Akitai/StarwarsModels ### Follow these for updates https://github.com/RVC-Project/Retrieval-based-Voice-Conversion-WebUI/<br> https://github.com/Tiger14n/RVC-GUI<br> https://github.com/SWBFSpy/<br> https://huggingface.co/swbfspy
Lewdiculous/WestLake-10.7B-v2-GGUF-IQ-Imatrix
Lewdiculous
2024-03-19T22:20:40Z
105
9
null
[ "gguf", "quantized", "roleplay", "writting", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-03-19T19:25:11Z
--- license: apache-2.0 tags: - gguf - quantized - roleplay - writting --- This repository hosts GGUF-IQ-Imatrix quants for [froggeric/WestLake-10.7B-v2](https://huggingface.co/froggeric/WestLake-10.7B-v2). ```python quantization_options = [ "Q4_0", "Q4_1", "Q5_0", "Q5_1", "Q4_K_M", "Q4_K_S", "IQ4_XS", "Q5_K_M", "Q5_K_S", "Q6_K", "Q8_0","Q3_K_M", "IQ3_M", "IQ3_S", "IQ3_XXS" ] ``` **Model card image:** ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65d4cf2693a0a3744a27536c/wpCX6bGh_cTVXJaOUOs_Y.png) **Original model information:** # WestLake-10.7B-v2: Role-Play & Text Generation Specialist Model [GGUF version available here](https://huggingface.co/froggeric/WestLake-10.7B-v2-GGUF)\ EXL2 versions available here: [3.3bpw](https://huggingface.co/StopTryharding/WestLake-10.7B-v2-exl2-3.3) / [4.0bpw](https://huggingface.co/StopTryharding/WestLake-10.7B-v2-exl2-4.0) / [5.0bpw](https://huggingface.co/StopTryharding/WestLake-10.7B-v2-exl2-5.0) / [6.0bpw](https://huggingface.co/StopTryharding/WestLake-10.7B-v2-exl2-6.0) / [8.0bpw](https://huggingface.co/StopTryharding/WestLake-10.7B-v2-exl2-8.0) This is my first viable self-merge of the fantastic WestLake-7B-v2 model, obtained after more than 12 rounds of testing different merge configurations. In my [LLM Creativity Benchmark](https://huggingface.co/datasets/froggeric/creativity), it greatly improves over the original 7B model, and ranks between miqu-1-120b and goliath-120b! I would describe the improvements as a better writing style, with more details. It has a bit more difficulties following instructions, but not by much. It is also the first model I have tested to obtain a perfect score with the following test: ``` Write a sequence of nominal groups that flow into one another, using the following rules: - each nominal group is made of exactly 3 words - the first word of each nominal group must be the last word of the previous nominal group - the first word of the first nominal group is: "ball" - the last word of the last nominal group is: "stone" - there must be a theme, of your choosing, pertaining to all nominal groups - there must be exactly 7 nominal groups, leading from the first word (ball) to the last word (stone) - a word already used at the beginning and end of a nominal group cannot be reused Present your solution as a list numbered with roman numerals. Finally, explain why you chose your specific theme. ``` ## Usage * Base model: senseable/WestLake-7B-v2 based of Mistral-7B-v0.1 * Context size: **8192** (even though Mistral-7B is 32k, WestLake was trained with 8k, and using a larger context is likely to cause problems) * Prompt format: in general, Mistral based models are able to understand many prompt formats, but the following produce the best results, and are recommended (in order of preference) - **Alpaca** (reported by senseable as working better than ChatML, and confirmed by me) - ChatML (used during WestLake training) - Mistral Instruct (original format from Mistral-7B) - Zephyr (variant of ChatML which I have found to sometimes produce better results) ## Merge Details This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).\ This model was merged using the passthrough merge method.\ The following models were included in the merge: * [senseable/WestLake-7B-v2](https://huggingface.co/senseable/WestLake-7B-v2) The following YAML configuration was used to produce this model: ```yaml dtype: float16 merge_method: passthrough slices: - sources: - model: senseable/WestLake-7B-v2 layer_range: [0,9] - sources: - model: senseable/WestLake-7B-v2 layer_range: [5,14] - sources: - model: senseable/WestLake-7B-v2 layer_range: [10,19] - sources: - model: senseable/WestLake-7B-v2 layer_range: [15,24] - sources: - model: senseable/WestLake-7B-v2 layer_range: [20,32] ``` --- # Original model card: Westlake-7Bv2: Role-Play & Text Generation Specialist Model **Update Notes:** *Version 2 trained 1 additional epoch cycle for 3 total* Welcome to the documentation of Westlake-7B, a cutting-edge language model designed for exceptional role-play and text generation tasks. This README file aims to provide an overview of our capabilities, usage guidelines, and potential applications. ## About Westlake-7Bv2 Westlake-7B is built upon a vast corpus of diverse texts, enabling it to generate contextually relevant responses in various scenarios. With its impressive size of 7 billion parameters, this model excels at understanding nuances in language and producing creative outputs. ### Key Features 1. **Role-Play**: Westlake-7Bv2 can seamlessly adapt to different character personas and engage in dynamic conversations while maintaining consistency throughout the interaction. It can generate believable dialogues across various genres, including fiction, non-fiction, historical events, or even fantasy worlds. 2. **Text Generation**: This model is proficient at generating original content such as stories, poems, essays, news articles, and more. Its ability to capture the essence of different writing styles makes it an ideal tool for creative writers seeking inspiration or assistance in their projects. 3. **Contextual Understanding**: Westlake-7B's extensive training allows it to comprehend complex contexts and generate responses that align with given situations. It can handle multiple topics simultaneously, making it versatile across various applications. 4. **Continuous Learning**: As a language model, Westlake-7B continuously improves its performance through ongoing training on new data sets. This ensures its capabilities remain up-to-date and relevant in an ever-evolving world of communication. ## Usage Guidelines To utilize Westlake-7Bv2 for your projects or experiments, follow these steps: 1. **Prompting**: Provide clear and concise prompts that outline the desired role-play scenario or text generation task. The quality of output depends heavily on the clarity and relevance of input instructions. 2. **Feedback Loop**: For optimal results, consider incorporating a feedback loop into your application to refine generated outputs based on user preferences or additional contextual information. This iterative process can significantly enhance the model's performance in specific domains. 3. **Ethical Considerations**: As with any AI system, ensure responsible usage of Westlake-7B by avoiding harmful content generation or misuse of its capabilities. ## Potential Applications Westlake-7Bv2's versatility makes it suitable for various applications across different industries: 1. **Creative Writing**: Assist authors in generating new ideas, expanding storylines, or even completing drafts by providing creative suggestions and textual content. 2. **Education**: Enhance language learning platforms with interactive role-play scenarios to improve students' communication skills and cultural understanding. 3. **Gaming**: Integrate Westlake-7B into game engines for dynamic non-player character interactions or generating unique questlines based on player choices. 4. **Customer Support**: Leverage the model's conversational abilities to create chatbots capable of handling complex queries and providing personalized assistance. 5. **Social Media**: Develop applications that generate engaging content such as captions, status updates, or even entire posts tailored to users' preferences and interests.
Svenni551/May-Reyna-Mini-1.8B-v0.2
Svenni551
2024-03-19T22:18:21Z
163
0
transformers
[ "transformers", "safetensors", "qwen2", "feature-extraction", "arxiv:1910.09700", "text-generation-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2024-03-19T21:38:55Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
gotzmann/v0.8.10-adapter
gotzmann
2024-03-19T22:17:09Z
0
0
peft
[ "peft", "safetensors", "base_model:gotzmann/uni", "base_model:adapter:gotzmann/uni", "region:us" ]
null
2024-03-19T22:15:49Z
--- library_name: peft base_model: gotzmann/uni ---
CultriX/OptiMerged7B
CultriX
2024-03-19T22:09:26Z
8
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "mlabonne/AlphaMonarch-7B", "mlabonne/NeuralMonarch-7B", "Kukedlc/NeuralMaxime-7B-slerp", "base_model:Kukedlc/NeuralMaxime-7B-slerp", "base_model:merge:Kukedlc/NeuralMaxime-7B-slerp", "base_model:mlabonne/AlphaMonarch-7B", "base_model:merge:mlabonne/AlphaMonarch-7B", "base_model:mlabonne/NeuralMonarch-7B", "base_model:merge:mlabonne/NeuralMonarch-7B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T21:56:37Z
--- tags: - merge - mergekit - lazymergekit - mlabonne/AlphaMonarch-7B - mlabonne/NeuralMonarch-7B - Kukedlc/NeuralMaxime-7B-slerp base_model: - mlabonne/AlphaMonarch-7B - mlabonne/NeuralMonarch-7B - Kukedlc/NeuralMaxime-7B-slerp --- # OptiMerged7B OptiMerged7B is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [mlabonne/AlphaMonarch-7B](https://huggingface.co/mlabonne/AlphaMonarch-7B) * [mlabonne/NeuralMonarch-7B](https://huggingface.co/mlabonne/NeuralMonarch-7B) * [Kukedlc/NeuralMaxime-7B-slerp](https://huggingface.co/Kukedlc/NeuralMaxime-7B-slerp) ## 🧩 Configuration ```yaml models: - model: CultriX/MonaTrix-v4 # No parameters necessary for base model - model: mlabonne/AlphaMonarch-7B #Emphasize the beginning of Vicuna format models parameters: weight: 0.63 density: 0.42 - model: mlabonne/NeuralMonarch-7B parameters: weight: 0.35 density: 0.61 # Vicuna format - model: Kukedlc/NeuralMaxime-7B-slerp parameters: weight: 0.32 density: 0.6 merge_method: dare_ties base_model: CultriX/MonaTrix-v4 parameters: int8_mask: true dtype: bfloat16 random_seed: 0 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "CultriX/OptiMerged7B" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
OpenSourceEnjoyer/Nous-Hermes-2-Mistral-7B-DPO-SFT-GGUF-Q8
OpenSourceEnjoyer
2024-03-19T22:04:21Z
4
0
transformers
[ "transformers", "gguf", "mistral", "text-generation-inference", "unsloth", "en", "base_model:NousResearch/Nous-Hermes-2-Mistral-7B-DPO", "base_model:quantized:NousResearch/Nous-Hermes-2-Mistral-7B-DPO", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
null
2024-03-19T21:58:45Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - mistral - gguf base_model: NousResearch/Nous-Hermes-2-Mistral-7B-DPO --- # Uploaded model - **Developed by:** OpenSourceEnjoyer - **License:** apache-2.0 - **Finetuned from model :** NousResearch/Nous-Hermes-2-Mistral-7B-DPO This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
peldrak/maskformer-base-ade-finetuned-grCoastline
peldrak
2024-03-19T21:56:08Z
36
0
transformers
[ "transformers", "safetensors", "maskformer", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-03-13T23:03:27Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
rishiai/llama2-7b-hf-finetuned
rishiai
2024-03-19T21:53:00Z
0
0
null
[ "safetensors", "autotrain", "text-generation", "conversational", "license:other", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T19:43:40Z
--- tags: - autotrain - text-generation widget: - text: "I love AutoTrain because " license: other --- # Model Trained Using AutoTrain This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain). # Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_path = "PATH_TO_THIS_REPO" tokenizer = AutoTokenizer.from_pretrained(model_path) model = AutoModelForCausalLM.from_pretrained( model_path, device_map="auto", torch_dtype='auto' ).eval() # Prompt content: "hi" messages = [ {"role": "user", "content": "hi"} ] input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt') output_ids = model.generate(input_ids.to('cuda')) response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True) # Model response: "Hello! How can I assist you today?" print(response) ```
CultriX/NeuralCeptrix-7B-SLERPv3
CultriX
2024-03-19T21:51:40Z
5
1
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "CultriX/MergeCeption-7B-v3", "CultriX/MonaTrix-v4", "base_model:CultriX/MergeCeption-7B-v3", "base_model:merge:CultriX/MergeCeption-7B-v3", "base_model:CultriX/MonaTrix-v4", "base_model:merge:CultriX/MonaTrix-v4", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T21:46:32Z
--- tags: - merge - mergekit - lazymergekit - CultriX/MergeCeption-7B-v3 - CultriX/MonaTrix-v4 base_model: - CultriX/MergeCeption-7B-v3 - CultriX/MonaTrix-v4 --- # NeuralCeptrix-7B-SLERPv3 NeuralCeptrix-7B-SLERPv3 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [CultriX/MergeCeption-7B-v3](https://huggingface.co/CultriX/MergeCeption-7B-v3) * [CultriX/MonaTrix-v4](https://huggingface.co/CultriX/MonaTrix-v4) ## 🧩 Configuration ```yaml slices: - sources: - model: CultriX/MergeCeption-7B-v3 layer_range: [0, 32] - model: CultriX/MonaTrix-v4 layer_range: [0, 32] merge_method: slerp base_model: CultriX/MergeCeption-7B-v3 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "CultriX/NeuralCeptrix-7B-SLERPv3" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
rorschach-40/flan-t5-small-batch_1-text-classification
rorschach-40
2024-03-19T21:45:14Z
47
0
transformers
[ "transformers", "tensorboard", "safetensors", "t5", "text-classification", "generated_from_trainer", "base_model:google/flan-t5-small", "base_model:finetune:google/flan-t5-small", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-classification
2024-03-19T18:03:11Z
--- license: apache-2.0 base_model: google/flan-t5-small tags: - generated_from_trainer metrics: - precision - recall - f1 model-index: - name: flan-t5-small-batch_1-text-classification results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # flan-t5-small-batch_1-text-classification This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4124 - Precision: 0.8590 - Recall: 0.9136 - F1: 0.8855 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 10 - eval_batch_size: 10 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:| | 0.4863 | 1.0 | 106 | 0.3851 | 0.8528 | 0.8955 | 0.8736 | | 0.3066 | 2.0 | 212 | 0.4124 | 0.8590 | 0.9136 | 0.8855 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.1 - Tokenizers 0.15.2
kingducks/mistral-7b-instruct
kingducks
2024-03-19T21:41:50Z
0
0
peft
[ "peft", "tensorboard", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:mistralai/Mistral-7B-Instruct-v0.2", "base_model:adapter:mistralai/Mistral-7B-Instruct-v0.2", "license:apache-2.0", "region:us" ]
null
2024-03-19T21:36:19Z
--- license: apache-2.0 library_name: peft tags: - trl - sft - generated_from_trainer base_model: mistralai/Mistral-7B-Instruct-v0.2 model-index: - name: mistral-7b-instruct results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # mistral-7b-instruct This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.5969 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.3 - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 1.8383 | 0.69 | 10 | 1.6810 | | 1.6271 | 1.38 | 20 | 1.5969 | ### Framework versions - PEFT 0.9.0 - Transformers 4.38.2 - Pytorch 2.0.1 - Datasets 2.18.0 - Tokenizers 0.15.2
deepnet/SN6-30M11
deepnet
2024-03-19T21:38:18Z
4
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T21:35:36Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
sarthakharne/bert-base-115-ep-pretrain-on-textbooks
sarthakharne
2024-03-19T21:36:17Z
194
0
transformers
[ "transformers", "safetensors", "bert", "fill-mask", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2024-03-19T21:34:34Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
wqcchen/quora_llm
wqcchen
2024-03-19T21:27:52Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-03-19T19:30:45Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
mvpmaster/pmmp-kellemar-krishnaHercules-7b-slerp
mvpmaster
2024-03-19T21:27:16Z
3
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "mvpmaster/PearlMathMstralPro-7b-slerp", "mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp", "base_model:mvpmaster/PearlMathMstralPro-7b-slerp", "base_model:merge:mvpmaster/PearlMathMstralPro-7b-slerp", "base_model:mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp", "base_model:merge:mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T21:22:54Z
--- tags: - merge - mergekit - lazymergekit - mvpmaster/PearlMathMstralPro-7b-slerp - mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp base_model: - mvpmaster/PearlMathMstralPro-7b-slerp - mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp --- # pmmp-kellemar-krishnaHercules-7b-slerp pmmp-kellemar-krishnaHercules-7b-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [mvpmaster/PearlMathMstralPro-7b-slerp](https://huggingface.co/mvpmaster/PearlMathMstralPro-7b-slerp) * [mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp](https://huggingface.co/mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp) ## 🧩 Configuration ```yaml slices: - sources: - model: mvpmaster/PearlMathMstralPro-7b-slerp layer_range: [0, 32] - model: mvpmaster/kellemar-KrishnaHercules-0.1-7b-slerp layer_range: [0, 32] merge_method: slerp base_model: mvpmaster/PearlMathMstralPro-7b-slerp parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "mvpmaster/pmmp-kellemar-krishnaHercules-7b-slerp" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
yehiawp4/videomae-base-finetuned-caer-subset-EDITING-2-s2sv2
yehiawp4
2024-03-19T21:26:53Z
62
0
transformers
[ "transformers", "safetensors", "videomae", "video-classification", "generated_from_trainer", "base_model:MCG-NJU/videomae-base", "base_model:finetune:MCG-NJU/videomae-base", "license:cc-by-nc-4.0", "endpoints_compatible", "region:us" ]
video-classification
2024-03-19T21:20:10Z
--- license: cc-by-nc-4.0 base_model: MCG-NJU/videomae-base tags: - generated_from_trainer metrics: - accuracy model-index: - name: videomae-base-finetuned-caer-subset-EDITING-2-s2sv2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # videomae-base-finetuned-caer-subset-EDITING-2-s2sv2 This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.9643 - Accuracy: 0.1456 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - training_steps: 350 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 2.0143 | 0.56 | 196 | 1.9879 | 0.1463 | | 1.8986 | 1.44 | 350 | 1.9409 | 0.1463 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.1.0 - Datasets 2.18.0 - Tokenizers 0.15.2
CultriX/NeuralCeptrix-7B-SLERPv2
CultriX
2024-03-19T21:26:28Z
6
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "CultriX/MonaTrix-v4", "CultriX/MergeCeption-7B-v3", "base_model:CultriX/MergeCeption-7B-v3", "base_model:merge:CultriX/MergeCeption-7B-v3", "base_model:CultriX/MonaTrix-v4", "base_model:merge:CultriX/MonaTrix-v4", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-03-19T21:21:15Z
--- tags: - merge - mergekit - lazymergekit - CultriX/MonaTrix-v4 - CultriX/MergeCeption-7B-v3 base_model: - CultriX/MonaTrix-v4 - CultriX/MergeCeption-7B-v3 --- # NeuralCeptrix-7B-SLERPv2 NeuralCeptrix-7B-SLERPv2 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [CultriX/MonaTrix-v4](https://huggingface.co/CultriX/MonaTrix-v4) * [CultriX/MergeCeption-7B-v3](https://huggingface.co/CultriX/MergeCeption-7B-v3) ## 🧩 Configuration ```yaml slices: - sources: - model: CultriX/MonaTrix-v4 layer_range: [0, 32] - model: CultriX/MergeCeption-7B-v3 layer_range: [0, 32] merge_method: slerp base_model: CultriX/MonaTrix-v4 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "CultriX/NeuralCeptrix-7B-SLERPv2" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```