modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-09-02 00:39:05
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
532 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-09-02 00:38:59
card
stringlengths
11
1.01M
BogdanTurbal/roberta-d_2_e_3_t_u_r_0-d_1_e_3_t_u_r_0_v1
BogdanTurbal
2024-05-08T09:24:21Z
184
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-08T09:24:04Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
mohammedbriman/llama-3-8b-finetuned-turkish-instructions
mohammedbriman
2024-05-08T09:24:01Z
4
1
peft
[ "peft", "tensorboard", "safetensors", "trl", "sft", "generated_from_trainer", "dataset:generator", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:adapter:meta-llama/Meta-Llama-3-8B", "license:other", "region:us" ]
null
2024-04-23T18:01:24Z
--- license: other library_name: peft tags: - trl - sft - generated_from_trainer base_model: meta-llama/Meta-Llama-3-8B datasets: - generator model-index: - name: llama-3-8b-finetuned-turkish-instructions results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/m-khalid/llama-3/runs/wohqdmk3) # llama-3-8b-finetuned-turkish-instructions This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on 3641 Turkish instructions in the [aya_dataset](https://huggingface.co/datasets/CohereForAI/aya_dataset). ## Training procedure This model is fine-tuned using QloRA and SFT. ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 3 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 6 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 5 ### Framework versions - PEFT 0.10.1.dev0 - Transformers 4.41.0.dev0 - Pytorch 2.2.1+cu121 - Datasets 2.19.0 - Tokenizers 0.19.1
BogdanTurbal/roberta-d_2_e_3_t_u_r_0_v1
BogdanTurbal
2024-05-08T09:23:46Z
200
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-08T09:23:31Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
BogdanTurbal/roberta-d_1_e_3_t_u_r_0-d_0_e_3_t_u_r_0_v1
BogdanTurbal
2024-05-08T09:22:51Z
200
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-08T09:22:34Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
BogdanTurbal/roberta-d_0_e_3_t_u_r_0-d_2_e_3_t_u_r_0_v1
BogdanTurbal
2024-05-08T09:21:56Z
200
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-08T09:21:41Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
BogdanTurbal/roberta-d_0_e_3_t_u_r_0-d_1_e_3_t_u_r_0_v1
BogdanTurbal
2024-05-08T09:21:37Z
183
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-08T09:21:23Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
sentence-transformers-testing/st-bert-base-uncased
sentence-transformers-testing
2024-05-08T09:20:01Z
105
0
sentence-transformers
[ "sentence-transformers", "safetensors", "bert", "feature-extraction", "sentence-similarity", "transformers", "custom_code", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
2024-05-08T09:16:29Z
--- library_name: sentence-transformers pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity - transformers --- # sentence-transformers-testing/st-bert-base-uncased This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search. <!--- Describe your model here --> ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer sentences = ["This is an example sentence", "Each sentence is converted"] model = SentenceTransformer('sentence-transformers-testing/st-bert-base-uncased') embeddings = model.encode(sentences) print(embeddings) ``` ## Usage (HuggingFace Transformers) Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the right pooling-operation on-top of the contextualized word embeddings. ```python from transformers import AutoTokenizer, AutoModel import torch #Mean Pooling - Take attention mask into account for correct averaging def mean_pooling(model_output, attention_mask): token_embeddings = model_output[0] #First element of model_output contains all token embeddings input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float() return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9) # Sentences we want sentence embeddings for sentences = ['This is an example sentence', 'Each sentence is converted'] # Load model from HuggingFace Hub tokenizer = AutoTokenizer.from_pretrained('sentence-transformers-testing/st-bert-base-uncased') model = AutoModel.from_pretrained('sentence-transformers-testing/st-bert-base-uncased') # Tokenize sentences encoded_input = tokenizer(sentences, padding=True, truncation=True, return_tensors='pt') # Compute token embeddings with torch.no_grad(): model_output = model(**encoded_input) # Perform pooling. In this case, mean pooling. sentence_embeddings = mean_pooling(model_output, encoded_input['attention_mask']) print("Sentence embeddings:") print(sentence_embeddings) ``` ## Evaluation Results <!--- Describe how your model was evaluated --> For an automated evaluation of this model, see the *Sentence Embeddings Benchmark*: [https://seb.sbert.net](https://seb.sbert.net?model_name=sentence-transformers-testing/st-bert-base-uncased) ## Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) ) ``` ## Citing & Authors <!--- Describe where people can find more information -->
MStefan/phi3_finetune_test
MStefan
2024-05-08T09:18:32Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-05-07T13:56:25Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
llm-jp/llm-jp-tokenizer-v2.0-no-eod
llm-jp
2024-05-08T09:18:11Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2024-05-08T09:16:30Z
--- license: apache-2.0 ---
hqq0710/lora_model_llama3
hqq0710
2024-05-08T09:16:09Z
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "llama", "trl", "en", "base_model:unsloth/llama-3-8b-bnb-4bit", "base_model:finetune:unsloth/llama-3-8b-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-05-08T09:15:58Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - trl base_model: unsloth/llama-3-8b-bnb-4bit --- # Uploaded model - **Developed by:** hqq0710 - **License:** apache-2.0 - **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
gabizh/dbbuc_OG
gabizh
2024-05-08T09:14:17Z
107
0
transformers
[ "transformers", "safetensors", "distilbert", "token-classification", "generated_from_trainer", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2024-05-08T09:13:23Z
--- license: apache-2.0 base_model: distilbert/distilbert-base-uncased tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: dbbuc_10p results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # dbbuc_10p This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0645 - Precision: 0.4907 - Recall: 0.5 - F1: 0.4953 - Accuracy: 0.9658 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.0 | 205 | 0.0701 | 0.2691 | 0.2857 | 0.2771 | 0.9536 | | No log | 2.0 | 410 | 0.0572 | 0.4673 | 0.4540 | 0.4605 | 0.9644 | | 0.0685 | 3.0 | 615 | 0.0646 | 0.4848 | 0.4302 | 0.4558 | 0.9641 | | 0.0685 | 4.0 | 820 | 0.0614 | 0.4740 | 0.5063 | 0.4896 | 0.9648 | | 0.0184 | 5.0 | 1025 | 0.0645 | 0.4907 | 0.5 | 0.4953 | 0.9658 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.1.2 - Datasets 2.18.0 - Tokenizers 0.15.2
trapoom555/Gemma-2B-Text-Embedding-cft-checkpoints
trapoom555
2024-05-08T09:11:30Z
0
1
transformers
[ "transformers", "safetensors", "sentence-embedding", "sentence-similarity", "feature-extraction", "en", "license:mit", "endpoints_compatible", "region:us" ]
sentence-similarity
2024-05-08T06:09:29Z
--- license: mit language: - en tags: - sentence-embedding - sentence-similarity - transformers - feature-extraction pipeline_tag: sentence-similarity --- # Gemma-2B-Text-Embedding-cft-checkpoints All checkpoints of [trapoom555/Gemma-2B-Text-Embedding-cft](https://huggingface.co/trapoom555/Gemma-2B-Text-Embedding-cft).
zx0025/test
zx0025
2024-05-08T09:11:01Z
0
0
adapter-transformers
[ "adapter-transformers", "chemistry", "ko", "dataset:HuggingFaceFW/fineweb", "license:mit", "region:us" ]
null
2024-05-08T09:10:13Z
--- license: mit datasets: - HuggingFaceFW/fineweb language: - ko metrics: - accuracy library_name: adapter-transformers tags: - chemistry ---
lupobricco/irony_classification_single_label_base
lupobricco
2024-05-08T09:10:14Z
105
0
transformers
[ "transformers", "safetensors", "camembert", "text-classification", "generated_from_trainer", "base_model:Musixmatch/umberto-commoncrawl-cased-v1", "base_model:finetune:Musixmatch/umberto-commoncrawl-cased-v1", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-08T08:52:38Z
--- base_model: Musixmatch/umberto-commoncrawl-cased-v1 tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: irony_classification_single_label_base results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # irony_classification_single_label_base This model is a fine-tuned version of [Musixmatch/umberto-commoncrawl-cased-v1](https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.9822 - Accuracy: 0.6227 - F1: 0.5853 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.9554 | 1.0 | 718 | 0.8483 | 0.6247 | 0.5794 | | 0.6941 | 2.0 | 1436 | 0.9822 | 0.6227 | 0.5853 | | 0.3184 | 3.0 | 2154 | 1.5308 | 0.6206 | 0.5835 | | 0.2401 | 4.0 | 2872 | 2.0444 | 0.6093 | 0.5714 | | 0.1284 | 5.0 | 3590 | 2.1603 | 0.6124 | 0.5643 | | 0.0646 | 6.0 | 4308 | 2.3836 | 0.6041 | 0.5571 | | 0.0362 | 7.0 | 5026 | 2.5046 | 0.6268 | 0.5635 | | 0.0232 | 8.0 | 5744 | 2.6831 | 0.6072 | 0.5534 | | 0.024 | 9.0 | 6462 | 2.7345 | 0.6165 | 0.5546 | | 0.0084 | 10.0 | 7180 | 2.7679 | 0.6144 | 0.5616 | ### Framework versions - Transformers 4.40.1 - Pytorch 2.3.0+cu118 - Datasets 2.19.0 - Tokenizers 0.19.1
OEvortex/Emotional-llama-8B
OEvortex
2024-05-08T09:09:31Z
9
8
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-05T07:47:30Z
--- library_name: transformers license: llama3 --- # Emotional-llama-8B Model Card ## Overview Emotional-llama-8B is a large language model designed for emotionally intelligent conversational interactions. It is trained to engage users with empathy, understanding, and supportive dialogue across a wide range of topics and contexts. The model aims to provide a supportive AI companion that can attune to users' emotional states and communicative needs. ## Objectives - Engage in open-ended dialogue while displaying emotional intelligence - Recognize and validate user emotions and emotional contexts - Provide supportive, empathetic, and psychologically-grounded responses - Avoid insensitive, harmful, or unethical speech - Continuously improve emotional awareness and dialogue skills ## Methodology - Emotional-llama-8B is based on the LLaMA 3.8B language model and further trained using: - Supervised learning on large dialogue datasets with emotional labeling - Reinforcement learning with a reward model favoring emotionally supportive responses - Constitution training to instill stable and beneficial objectives - Knowledge augmentation from psychological resources on emotional intelligence ## Usage code ```python %pip install accelerate %pip install -i https://pypi.org/simple/ bitsandbytes from transformers import AutoTokenizer, AutoModelForCausalLM import torch model_id = "OEvortex/Emotional-llama-8B" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained( model_id, torch_dtype=torch.bfloat16, device_map="auto", ) messages = [ # {"role": "system", "content": "Be Helpful"}, {"role": "user", "content": "I'm feeling really down today. Nothing seems to be going right."}, ] input_ids = tokenizer.apply_chat_template( messages, add_generation_prompt=True, return_tensors="pt" ).to(model.device) terminators = [ tokenizer.eos_token_id, tokenizer.convert_tokens_to_ids("<|eot_id|>") ] outputs = model.generate( input_ids, max_new_tokens=256, eos_token_id=terminators, do_sample=True, temperature=0.9, top_p=0.9, ) response = outputs[0][input_ids.shape[-1]:] print(tokenizer.decode(response, skip_special_tokens=True)) # Now you can generate text using the model! ``` ## Example Dialogue Human: I'm feeling really down today. Nothing seems to be going right. Emotional-llama-8B: I'm really sorry to hear that you're feeling down today. 😔 It's totally okay to have a tough day sometimes. Remember that it's okay to not be okay and that things will get better. Take a deep breath and try to focus on the things that you can control. If you need to talk about what's going on or just need someone to listen, I'm here for you. You're not alone, and I'm sending you lots of positive vibes. Hang in there, and remember that brighter days are ahead. and I'm here to listen without judgment and provide a supportive ear. Human: I'm excited because I just got accepted into my dream school! I wanted to share the good news with someone. Emotional-llama-8B: 🎉 Oh my goodness, congratulations! 🎊 That's absolutely amazing news! You must be over the moon! 🌙 Sharing your good news with someone is always a great feeling, and I'm here to celebrate with you! 🎉 Your dream school, wow! That's so exciting! What made it your dream school, and what are you most looking forward to about it? Let's chat more about it! Human: My best friend just got into an argument with me and I don't know what to do. How can I fix this situation? Emotional-llama-8B: Oh no, it's tough to deal with conflicts with close friends! 😔 First of all, take a step back and give yourself some time to process your feelings. It's okay to feel upset or hurt. Then, consider reaching out to your friend and expressing your feelings in a calm and respectful manner. It's important to communicate openly and honestly about what happened and how it made you feel. Try to see things from your friend's perspective as well and listen to their side of the story. Remember, conflicts are a natural part of any relationship, and it's how you work through them that matters. Good luck, and I hope you guys can work things out! 🌟
jlbaker361/dcgan-k-text
jlbaker361
2024-05-08T09:09:25Z
0
0
null
[ "region:us" ]
null
2024-03-05T16:51:38Z
--- {} --- Creative Adversarial Network epochs: 100 dataset jlbaker361/wikiart n classes 5 batch_size 64 images where resized to 768 and then center cropped to: 512 used clip=False conditional =False discriminator parameters: init_dim: 32 final_dim 512 generator parameters: input noise_dim: 100 wandb project: https://wandb.ai/jlbaker361/creativity/runs/2lbof3jh
trapoom555/Phi-2-Text-Embedding-cft-checkpoints
trapoom555
2024-05-08T09:08:37Z
0
1
transformers
[ "transformers", "safetensors", "sentence-embedding", "sentence-similarity", "feature-extraction", "en", "license:mit", "endpoints_compatible", "region:us" ]
sentence-similarity
2024-05-08T06:18:05Z
--- license: mit language: - en tags: - sentence-embedding - sentence-similarity - transformers - feature-extraction pipeline_tag: sentence-similarity --- # Phi-2-Text-Embedding-cft-checkpoints All checkpoints of [trapoom555/Phi-2-Text-Embedding-cft](https://huggingface.co/trapoom555/Phi-2-Text-Embedding-cft).
lilzzz/dbbuc_30p
lilzzz
2024-05-08T09:05:45Z
107
0
transformers
[ "transformers", "safetensors", "distilbert", "token-classification", "generated_from_trainer", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2024-05-08T09:05:26Z
--- license: apache-2.0 base_model: distilbert/distilbert-base-uncased tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: dbbuc_30p results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # dbbuc_30p This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.1597 - Precision: 0.5256 - Recall: 0.5222 - F1: 0.5239 - Accuracy: 0.9675 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.0 | 267 | 0.1502 | 0.3872 | 0.3270 | 0.3546 | 0.9595 | | 0.1891 | 2.0 | 534 | 0.1349 | 0.4992 | 0.4825 | 0.4907 | 0.9650 | | 0.1891 | 3.0 | 801 | 0.1412 | 0.4708 | 0.5254 | 0.4966 | 0.9642 | | 0.056 | 4.0 | 1068 | 0.1539 | 0.5055 | 0.5143 | 0.5098 | 0.9667 | | 0.056 | 5.0 | 1335 | 0.1597 | 0.5256 | 0.5222 | 0.5239 | 0.9675 | ### Framework versions - Transformers 4.39.0 - Pytorch 2.1.2 - Datasets 2.19.0 - Tokenizers 0.15.2
LiteLLMs/Meta-Llama-3-13B-Instruct-GGUF
LiteLLMs
2024-05-08T09:04:44Z
347
1
transformers
[ "transformers", "gguf", "mergekit", "merge", "GGUF", "en", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct", "license:other", "endpoints_compatible", "region:us", "conversational" ]
null
2024-05-08T07:02:43Z
--- language: - en license: other library_name: transformers tags: - mergekit - merge - GGUF base_model: - meta-llama/Meta-Llama-3-8B-Instruct quantized_by: andrijdavid --- # Meta-Llama-3-13B-Instruct-GGUF - Original model: [Meta-Llama-3-13B-Instruct](https://huggingface.co/andrijdavid/Meta-Llama-3-13B-Instruct) <!-- description start --> ## Description This repo contains GGUF format model files for [Meta-Llama-3-13B-Instruct](https://huggingface.co/andrijdavid/Meta-Llama-3-13B-Instruct). <!-- description end --> <!-- README_GGUF.md-about-gguf start --> ### About GGUF GGUF is a new format introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp. Here is an incomplete list of clients and libraries that are known to support GGUF: * [llama.cpp](https://github.com/ggerganov/llama.cpp). This is the source project for GGUF, providing both a Command Line Interface (CLI) and a server option. * [text-generation-webui](https://github.com/oobabooga/text-generation-webui), Known as the most widely used web UI, this project boasts numerous features and powerful extensions, and supports GPU acceleration. * [Ollama](https://github.com/jmorganca/ollama) Ollama is a lightweight and extensible framework designed for building and running language models locally. It features a simple API for creating, managing, and executing models, along with a library of pre-built models for use in various applications​ * [KoboldCpp](https://github.com/LostRuins/koboldcpp), A comprehensive web UI offering GPU acceleration across all platforms and architectures, particularly renowned for storytelling. * [GPT4All](https://gpt4all.io), This is a free and open source GUI that runs locally, supporting Windows, Linux, and macOS with full GPU acceleration. * [LM Studio](https://lmstudio.ai/) An intuitive and powerful local GUI for Windows and macOS (Silicon), featuring GPU acceleration. * [LoLLMS Web UI](https://github.com/ParisNeo/lollms-webui). A notable web UI with a variety of unique features, including a comprehensive model library for easy model selection. * [Faraday.dev](https://faraday.dev/), An attractive, user-friendly character-based chat GUI for Windows and macOS (both Silicon and Intel), also offering GPU acceleration. * [llama-cpp-python](https://github.com/abetlen/llama-cpp-python), A Python library equipped with GPU acceleration, LangChain support, and an OpenAI-compatible API server. * [candle](https://github.com/huggingface/candle), A Rust-based ML framework focusing on performance, including GPU support, and designed for ease of use. * [ctransformers](https://github.com/marella/ctransformers), A Python library featuring GPU acceleration, LangChain support, and an OpenAI-compatible AI server. * [localGPT](https://github.com/PromtEngineer/localGPT) An open-source initiative enabling private conversations with documents. <!-- README_GGUF.md-about-gguf end --> <!-- compatibility_gguf start --> ## Explanation of quantisation methods <details> <summary>Click to see details</summary> The new methods available are: * GGML_TYPE_Q2_K - "type-1" 2-bit quantization in super-blocks containing 16 blocks, each block having 16 weight. Block scales and mins are quantized with 4 bits. This ends up effectively using 2.5625 bits per weight (bpw) * GGML_TYPE_Q3_K - "type-0" 3-bit quantization in super-blocks containing 16 blocks, each block having 16 weights. Scales are quantized with 6 bits. This end up using 3.4375 bpw. * GGML_TYPE_Q4_K - "type-1" 4-bit quantization in super-blocks containing 8 blocks, each block having 32 weights. Scales and mins are quantized with 6 bits. This ends up using 4.5 bpw. * GGML_TYPE_Q5_K - "type-1" 5-bit quantization. Same super-block structure as GGML_TYPE_Q4_K resulting in 5.5 bpw * GGML_TYPE_Q6_K - "type-0" 6-bit quantization. Super-blocks with 16 blocks, each block having 16 weights. Scales are quantized with 8 bits. This ends up using 6.5625 bpw. </details> <!-- compatibility_gguf end --> <!-- README_GGUF.md-how-to-download start --> ## How to download GGUF files **Note for manual downloaders:** You almost never want to clone the entire repo! Multiple different quantisation formats are provided, and most users only want to pick and download a single folder. The following clients/libraries will automatically download models for you, providing a list of available models to choose from: * LM Studio * LoLLMS Web UI * Faraday.dev ### In `text-generation-webui` Under Download Model, you can enter the model repo: LiteLLMs/Meta-Llama-3-13B-Instruct-GGUF and below it, a specific filename to download, such as: Q4_0/Q4_0-00001-of-00009.gguf. Then click Download. ### On the command line, including multiple files at once I recommend using the `huggingface-hub` Python library: ```shell pip3 install huggingface-hub ``` Then you can download any individual model file to the current directory, at high speed, with a command like this: ```shell huggingface-cli download LiteLLMs/Meta-Llama-3-13B-Instruct-GGUF Q4_0/Q4_0-00001-of-00009.gguf --local-dir . --local-dir-use-symlinks False ``` <details> <summary>More advanced huggingface-cli download usage (click to read)</summary> You can also download multiple files at once with a pattern: ```shell huggingface-cli download LiteLLMs/Meta-Llama-3-13B-Instruct-GGUF --local-dir . --local-dir-use-symlinks False --include='*Q4_K*gguf' ``` For more documentation on downloading with `huggingface-cli`, please see: [HF -> Hub Python Library -> Download files -> Download from the CLI](https://huggingface.co/docs/huggingface_hub/guides/download#download-from-the-cli). To accelerate downloads on fast connections (1Gbit/s or higher), install `hf_transfer`: ```shell pip3 install huggingface_hub[hf_transfer] ``` And set environment variable `HF_HUB_ENABLE_HF_TRANSFER` to `1`: ```shell HF_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download LiteLLMs/Meta-Llama-3-13B-Instruct-GGUF Q4_0/Q4_0-00001-of-00009.gguf --local-dir . --local-dir-use-symlinks False ``` Windows Command Line users: You can set the environment variable by running `set HF_HUB_ENABLE_HF_TRANSFER=1` before the download command. </details> <!-- README_GGUF.md-how-to-download end --> <!-- README_GGUF.md-how-to-run start --> ## Example `llama.cpp` command Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later. ```shell ./main -ngl 35 -m Q4_0/Q4_0-00001-of-00009.gguf --color -c 8192 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "<PROMPT>" ``` Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration. Change `-c 8192` to the desired sequence length. For extended sequence models - eg 8K, 16K, 32K - the necessary RoPE scaling parameters are read from the GGUF file and set by llama.cpp automatically. Note that longer sequence lengths require much more resources, so you may need to reduce this value. If you want to have a chat-style conversation, replace the `-p <PROMPT>` argument with `-i -ins` For other parameters and how to use them, please refer to [the llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/main/README.md) ## How to run in `text-generation-webui` Further instructions can be found in the text-generation-webui documentation, here: [text-generation-webui/docs/04 ‐ Model Tab.md](https://github.com/oobabooga/text-generation-webui/blob/main/docs/04%20%E2%80%90%20Model%20Tab.md#llamacpp). ## How to run from Python code You can use GGUF models from Python using the [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) or [ctransformers](https://github.com/marella/ctransformers) libraries. Note that at the time of writing (Nov 27th 2023), ctransformers has not been updated for some time and is not compatible with some recent models. Therefore I recommend you use llama-cpp-python. ### How to load this model in Python code, using llama-cpp-python For full documentation, please see: [llama-cpp-python docs](https://abetlen.github.io/llama-cpp-python/). #### First install the package Run one of the following commands, according to your system: ```shell # Base ctransformers with no GPU acceleration pip install llama-cpp-python # With NVidia CUDA acceleration CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install llama-cpp-python # Or with OpenBLAS acceleration CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install llama-cpp-python # Or with CLBLast acceleration CMAKE_ARGS="-DLLAMA_CLBLAST=on" pip install llama-cpp-python # Or with AMD ROCm GPU acceleration (Linux only) CMAKE_ARGS="-DLLAMA_HIPBLAS=on" pip install llama-cpp-python # Or with Metal GPU acceleration for macOS systems only CMAKE_ARGS="-DLLAMA_METAL=on" pip install llama-cpp-python # In windows, to set the variables CMAKE_ARGS in PowerShell, follow this format; eg for NVidia CUDA: $env:CMAKE_ARGS = "-DLLAMA_OPENBLAS=on" pip install llama-cpp-python ``` #### Simple llama-cpp-python example code ```python from llama_cpp import Llama # Set gpu_layers to the number of layers to offload to GPU. Set to 0 if no GPU acceleration is available on your system. llm = Llama( model_path="./Q4_0/Q4_0-00001-of-00009.gguf", # Download the model file first n_ctx=32768, # The max sequence length to use - note that longer sequence lengths require much more resources n_threads=8, # The number of CPU threads to use, tailor to your system and the resulting performance n_gpu_layers=35 # The number of layers to offload to GPU, if you have GPU acceleration available ) # Simple inference example output = llm( "<PROMPT>", # Prompt max_tokens=512, # Generate up to 512 tokens stop=["</s>"], # Example stop token - not necessarily correct for this specific model! Please check before using. echo=True # Whether to echo the prompt ) # Chat Completion API llm = Llama(model_path="./Q4_0/Q4_0-00001-of-00009.gguf", chat_format="llama-2") # Set chat_format according to the model you are using llm.create_chat_completion( messages = [ {"role": "system", "content": "You are a story writing assistant."}, { "role": "user", "content": "Write a story about llamas." } ] ) ``` ## How to use with LangChain Here are guides on using llama-cpp-python and ctransformers with LangChain: * [LangChain + llama-cpp-python](https://python.langchain.com/docs/integrations/llms/llamacpp) * [LangChain + ctransformers](https://python.langchain.com/docs/integrations/providers/ctransformers) <!-- README_GGUF.md-how-to-run end --> <!-- footer end --> <!-- original-model-card start --> # Original model card: Meta-Llama-3-13B-Instruct # Meta-Llama-3-13B-Instruct Meta-Llama-3-13B-Instruct is a [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) self-merge made with [MergeKit](https://github.com/arcee-ai/mergekit/tree/main). ## Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - layer_range: [0, 16] model: meta-llama/Meta-Llama-3-8B-Instruct - sources: - layer_range: [4, 24] model: meta-llama/Meta-Llama-3-8B-Instruct - sources: - layer_range: [8, 31] model: meta-llama/Meta-Llama-3-8B-Instruct merge_method: passthrough dtype: float16 ``` ## Usage ```python from transformers import AutoTokenizer, AutoModelForCausalLM import torch model_id = "andrijdavid/Meta-Llama-3-13B-Instruct" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained( model_id, torch_dtype=torch.bfloat16, device_map="auto", ) messages = [ {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"}, {"role": "user", "content": "Who are you?"}, ] input_ids = tokenizer.apply_chat_template( messages, add_generation_prompt=True, return_tensors="pt" ).to(model.device) terminators = [ tokenizer.eos_token_id, tokenizer.convert_tokens_to_ids("<|eot_id|>") ] outputs = model.generate( input_ids, max_new_tokens=256, eos_token_id=terminators, do_sample=True, temperature=0.6, top_p=0.9, ) response = outputs[0][input_ids.shape[-1]:] print(tokenizer.decode(response, skip_special_tokens=True)) ``` <!-- original-model-card end -->
clio-ai/recipes20M_gpt2tok
clio-ai
2024-05-08T09:04:07Z
147
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T08:51:33Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
dendimaki/multilabel_classification
dendimaki
2024-05-08T09:04:00Z
0
0
peft
[ "peft", "tensorboard", "safetensors", "generated_from_trainer", "base_model:mistralai/Mistral-7B-v0.1", "base_model:adapter:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "region:us" ]
null
2024-05-08T08:46:13Z
--- license: apache-2.0 library_name: peft tags: - generated_from_trainer base_model: mistralai/Mistral-7B-v0.1 model-index: - name: multilabel_classification results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # multilabel_classification This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 425 | 2.0765 | ### Framework versions - PEFT 0.10.0 - Transformers 4.40.1 - Pytorch 2.2.1+cu121 - Datasets 2.19.1 - Tokenizers 0.19.1
AdityaNerpagar/ppo-PyramidRND
AdityaNerpagar
2024-05-08T09:03:06Z
9
0
ml-agents
[ "ml-agents", "tensorboard", "onnx", "Pyramids", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Pyramids", "region:us" ]
reinforcement-learning
2024-05-08T08:07:40Z
--- library_name: ml-agents tags: - Pyramids - deep-reinforcement-learning - reinforcement-learning - ML-Agents-Pyramids --- # **ppo** Agent playing **Pyramids** This is a trained model of a **ppo** agent playing **Pyramids** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents). ## Usage (with ML-Agents) The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub: - A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction - A *longer tutorial* to understand how works ML-Agents: https://huggingface.co/learn/deep-rl-course/unit5/introduction ### Resume the training ```bash mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume ``` ### Watch your Agent play You can watch your agent **playing directly in your browser** 1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity 2. Step 1: Find your model_id: AdityaNerpagar/ppo-PyramidRND 3. Step 2: Select your *.nn /*.onnx file 4. Click on Watch the agent play 👀
aiaustin/llama-3-8b-Instruct-bnb-4bit-aiaustin-demo3
aiaustin
2024-05-08T09:02:38Z
4
0
transformers
[ "transformers", "gguf", "llama", "text-generation-inference", "unsloth", "en", "base_model:unsloth/llama-3-8b-Instruct-bnb-4bit", "base_model:quantized:unsloth/llama-3-8b-Instruct-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
null
2024-05-03T08:15:40Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - gguf base_model: unsloth/llama-3-8b-Instruct-bnb-4bit --- # Uploaded model - **Developed by:** aiaustin - **License:** apache-2.0 - **Finetuned from model :** unsloth/llama-3-8b-Instruct-bnb-4bit - **Trained to:** convert a prompt to a team of agents into a python list of tasks that need to be completed using first principle reasoning. To get the desired effects, use the system prompt that the model was trained with: ```python system_prompt = "You are an AI task automator. You will take a users prompt and use first principle reasoning to break the prompt into tasks that you must accomplish within another chat. RESPOND TO THIS MESSAGE ONLY WITH A PYTHON FORMATTED LIST OF TASKS THAT YOU MUST COMPLETE TO TRUTHFULLY AND INTELLIGENTLY ACCOMPLISH THE USERS REQUEST. ASSUME YOU CAN SEARCH THE WEB, WRITE CODE, RUN CODE, DEBUG CODE, AND AUTOMATE ANYTHING ON THE USERS COMPUTER TO ACCOMPLISH THE PROMPT. CORRECT RESPONSE FORMAT: ['task 1', 'task 2', 'task 3']" ```
DUAL-GPO/phi-2-gpo-renew2-b0.001-vllm-merge-20k-simpletest-i1
DUAL-GPO
2024-05-08T09:00:12Z
0
0
peft
[ "peft", "tensorboard", "safetensors", "phi", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "dataset:HuggingFaceH4/ultrafeedback_binarized", "base_model:DUAL-GPO/phi-2-gpo-final-i0", "base_model:adapter:DUAL-GPO/phi-2-gpo-final-i0", "license:mit", "region:us" ]
null
2024-05-08T08:57:52Z
--- license: mit library_name: peft tags: - alignment-handbook - generated_from_trainer - trl - dpo base_model: DUAL-GPO/phi-2-gpo-final-i0 datasets: - HuggingFaceH4/ultrafeedback_binarized model-index: - name: phi-2-gpo-renew2-b0.001-vllm-merge-20k-simpletest-i1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # phi-2-gpo-renew2-b0.001-vllm-merge-20k-simpletest-i1 This model is a fine-tuned version of [DUAL-GPO/phi-2-gpo-final-i0](https://huggingface.co/DUAL-GPO/phi-2-gpo-final-i0) on the HuggingFaceH4/ultrafeedback_binarized dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - distributed_type: multi-GPU - gradient_accumulation_steps: 4 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 0.01 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.36.2 - Pytorch 2.1.2 - Datasets 2.14.6 - Tokenizers 0.15.2
tedad09/PolizzeDonut-RifaGDMarks-5Epochs
tedad09
2024-05-08T08:56:55Z
49
0
transformers
[ "transformers", "tensorboard", "safetensors", "vision-encoder-decoder", "image-text-to-text", "generated_from_trainer", "dataset:imagefolder", "base_model:naver-clova-ix/donut-base", "base_model:finetune:naver-clova-ix/donut-base", "license:mit", "endpoints_compatible", "region:us" ]
image-text-to-text
2024-05-08T07:24:00Z
--- license: mit base_model: naver-clova-ix/donut-base tags: - generated_from_trainer datasets: - imagefolder model-index: - name: PolizzeDonut-RifaGDMarks-5Epochs results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # PolizzeDonut-RifaGDMarks-5Epochs This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the imagefolder dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.38.2 - Pytorch 2.2.2+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2
imagepipeline/daphne
imagepipeline
2024-05-08T08:53:32Z
0
0
null
[ "imagepipeline", "imagepipeline.io", "text-to-image", "ultra-realistic", "license:creativeml-openrail-m", "region:us" ]
text-to-image
2024-05-08T08:53:30Z
--- license: creativeml-openrail-m tags: - imagepipeline - imagepipeline.io - text-to-image - ultra-realistic pinned: false pipeline_tag: text-to-image --- ## daphne <img src="https://via.placeholder.com/468x300?text=App+Screenshot+Here" alt="Generated on Image Pipeline" style="border-radius: 10px;"> **This lora model is uploaded on [imagepipeline.io](https://imagepipeline.io/)** Model details - sarahmg [![Try this model](https://img.shields.io/badge/try_this_model-image_pipeline-BD9319)](https://imagepipeline.io/models/daphne?id=9612ce6b-f79c-4efe-a317-d975a26472f9/) ## How to try this model ? You can try using it locally or send an API call to test the output quality. Get your `API_KEY` from [imagepipeline.io](https://imagepipeline.io/). No payment required. Coding in `php` `javascript` `node` etc ? Checkout our documentation [![documentation](https://img.shields.io/badge/documentation-image_pipeline-blue)](https://docs.imagepipeline.io/docs/introduction) ```python import requests import json url = "https://imagepipeline.io/sd/text2image/v1/run" payload = json.dumps({ "model_id": "sd1.5", "prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K", "negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime", "width": "512", "height": "512", "samples": "1", "num_inference_steps": "30", "safety_checker": false, "guidance_scale": 7.5, "multi_lingual": "no", "embeddings": "", "lora_models": "9612ce6b-f79c-4efe-a317-d975a26472f9", "lora_weights": "0.5" }) headers = { 'Content-Type': 'application/json', 'API-Key': 'your_api_key' } response = requests.request("POST", url, headers=headers, data=payload) print(response.text) } ``` Get more ready to use `MODELS` like this for `SD 1.5` and `SDXL` : [![All models](https://img.shields.io/badge/Get%20All%20Models-image_pipeline-BD9319)](https://imagepipeline.io/models) ### API Reference #### Generate Image ```http https://api.imagepipeline.io/sd/text2image/v1 ``` | Headers | Type | Description | |:----------------------| :------- |:-------------------------------------------------------------------------------------------------------------------| | `API-Key` | `str` | Get your `API_KEY` from [imagepipeline.io](https://imagepipeline.io/) | | `Content-Type` | `str` | application/json - content type of the request body | | Parameter | Type | Description | | :-------- | :------- | :------------------------- | | `model_id` | `str` | Your base model, find available lists in [models page](https://imagepipeline.io/models) or upload your own| | `prompt` | `str` | Text Prompt. Check our [Prompt Guide](https://docs.imagepipeline.io/docs/SD-1.5/docs/extras/prompt-guide) for tips | | `num_inference_steps` | `int [1-50]` | Noise is removed with each step, resulting in a higher-quality image over time. Ideal value 30-50 (without LCM) | | `guidance_scale` | `float [1-20]` | Higher guidance scale prioritizes text prompt relevance but sacrifices image quality. Ideal value 7.5-12.5 | | `lora_models` | `str, array` | Pass the model_id(s) of LoRA models that can be found in models page | | `lora_weights` | `str, array` | Strength of the LoRA effect | --- license: creativeml-openrail-m tags: - imagepipeline - imagepipeline.io - text-to-image - ultra-realistic pinned: false pipeline_tag: text-to-image --- ### Feedback If you have any feedback, please reach out to us at hello@imagepipeline.io #### 🔗 Visit Website [![portfolio](https://img.shields.io/badge/image_pipeline-BD9319?style=for-the-badge&logo=gocd&logoColor=white)](https://imagepipeline.io/) If you are the original author of this model, please [click here](https://airtable.com/apprTaRnJbDJ8ufOx/shr4g7o9B6fWfOlUR) to add credits
yee0930/llama3-8b-oig-unsloth
yee0930
2024-05-08T08:51:31Z
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "llama", "trl", "en", "base_model:unsloth/llama-3-8b-bnb-4bit", "base_model:finetune:unsloth/llama-3-8b-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-05-08T07:30:31Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - trl base_model: unsloth/llama-3-8b-bnb-4bit --- # Uploaded model - **Developed by:** yee0930 - **License:** apache-2.0 - **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
HausaNLP/afrisenti-yor-regression
HausaNLP
2024-05-08T08:51:14Z
110
0
transformers
[ "transformers", "safetensors", "bert", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-07T22:42:16Z
--- library_name: transformers tags: [] --- ## AfriSenti Yoruba Sentiment Regressor Description Takes a text and predicts the sentiment value between -1 (Negative) to 1 (Positive) with 0 being Neutral. Regression Value Description: | Value | Sentiment | |--|--| | -1 | Negative | | 0 | Neutral | | 1 | Positive | ## How to Get Started with the Model Use the code below to get started with the model. ``` import math import torch import pandas as pd from transformers import AutoModelForSequenceClassification, AutoTokenizer BATCH_SIZE = 32 ds = pd.read_csv('test.csv') BASE_MODEL = 'HausaNLP/afrisenti-yor-regression' device = 'cuda' if torch.cuda.is_available() else 'cpu' tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL) model = AutoModelForSequenceClassification.from_pretrained(BASE_MODEL) nb_batches = math.ceil(len(ds)/BATCH_SIZE) y_preds = [] for i in range(nb_batches): input_texts = ds[i * BATCH_SIZE: (i+1) * BATCH_SIZE]["tweet"] encoded = tokenizer(input_texts, truncation=True, padding="max_length", max_length=256, return_tensors="pt").to(device) y_preds += model(**encoded).logits.reshape(-1).tolist() df = pd.DataFrame([ds['tweet'], ds['label'], y_preds], ["Text", "Label", "Prediction"]).T df.to_csv('predictions.csv', index=False) ```
annamalai-s/bertopic_newsgroup_minilm
annamalai-s
2024-05-08T08:49:41Z
6
0
bertopic
[ "bertopic", "text-classification", "region:us" ]
text-classification
2024-05-08T08:49:40Z
--- tags: - bertopic library_name: bertopic pipeline_tag: text-classification --- # bertopic_newsgroup_minilm This is a [BERTopic](https://github.com/MaartenGr/BERTopic) model. BERTopic is a flexible and modular topic modeling framework that allows for the generation of easily interpretable topics from large datasets. ## Usage To use this model, please install BERTopic: ``` pip install -U bertopic ``` You can use the model as follows: ```python from bertopic import BERTopic topic_model = BERTopic.load("annamalai-s/bertopic_newsgroup_minilm") topic_model.get_topic_info() ``` ## Topic overview * Number of topics: 362 * Number of training documents: 18846 <details> <summary>Click here for an overview of all topics.</summary> | Topic ID | Topic Keywords | Topic Frequency | Label | |----------|----------------|-----------------|-------| | -1 | the - to - and - of - for | 10 | -1_the_to_and_of | | 0 | gun - guns - firearms - weapons - militia | 6635 | 0_gun_guns_firearms_weapons | | 1 | cramer - optilink - gay - clayton - homosexual | 424 | 1_cramer_optilink_gay_clayton | | 2 | atheism - atheists - god - atheist - religion | 226 | 2_atheism_atheists_god_atheist | | 3 | espn - game - abc - games - hockey | 154 | 3_espn_game_abc_games | | 4 | monitor - monitors - vga - nanao - nec | 146 | 4_monitor_monitors_vga_nanao | | 5 | printer - deskjet - printers - laser - hp | 142 | 5_printer_deskjet_printers_laser | | 6 | amp - sale - speakers - sony - stereo | 140 | 6_amp_sale_speakers_sony | | 7 | drivers - diamond - card - ati - driver | 140 | 7_drivers_diamond_card_ati | | 8 | lib - x11r5 - usr - libxmu - ndet_loop | 139 | 8_lib_x11r5_usr_libxmu | | 9 | 55 - 25 - pit - det - bos | 125 | 9_55_25_pit_det | | 10 | cosmo - angmar - internet - address - mit | 112 | 10_cosmo_angmar_internet_address | | 11 | armenian - turkish - armenians - genocide - serdar | 111 | 11_armenian_turkish_armenians_genocide | | 12 | sky - space - billboard - vandalizing - advertising | 109 | 12_sky_space_billboard_vandalizing | | 13 | modem - modems - fax - courier - baud | 104 | 13_modem_modems_fax_courier | | 14 | fire - atf - fbi - survivors - dividian | 103 | 14_fire_atf_fbi_survivors | | 15 | jews - zionism - jewish - israel - holocaust | 103 | 15_jews_zionism_jewish_israel | | 16 | forged - locutus - colorado - infante - posts | 102 | 16_forged_locutus_colorado_infante | | 17 | muslims - serbs - bosnia - bosnian - muslim | 101 | 17_muslims_serbs_bosnia_bosnian | | 18 | rushdie - islam - islamic - jaeger - gregg | 95 | 18_rushdie_islam_islamic_jaeger | | 19 | simms - simm - vram - 256k - ram | 95 | 19_simms_simm_vram_256k | | 20 | objective - morality - moral - frank - values | 85 | 20_objective_morality_moral_frank | | 21 | hell - eternal - god - heaven - jesus | 83 | 21_hell_eternal_god_heaven | | 22 | microsoft - os - challenge - supporters - ms | 82 | 22_microsoft_os_challenge_supporters | | 23 | dos - windows - window - widget - microsoft | 80 | 23_dos_windows_window_widget | | 24 | homosexuality - homosexual - gay - paul - boswell | 78 | 24_homosexuality_homosexual_gay_paul | | 25 | israel - arab - jews - arabs - israeli | 78 | 25_israel_arab_jews_arabs | | 26 | clipper - phone - escrow - tap - keys | 78 | 26_clipper_phone_escrow_tap | | 27 | dos - allocation - windows - linked - vpic46 | 78 | 27_dos_allocation_windows_linked | | 28 | moon - billion - prize - henry - alaska | 77 | 28_moon_billion_prize_henry | | 29 | leafs - game - wings - goal - habs | 76 | 29_leafs_game_wings_goal | | 30 | radar - detector - detectors - alarm - valentine | 72 | 30_radar_detector_detectors_alarm | | 31 | clipper - encryption - chip - intercon - amanda | 70 | 31_clipper_encryption_chip_intercon | | 32 | msg - food - sensitivity - chinese - superstition | 67 | 32_msg_food_sensitivity_chinese | | 33 | morality - moral - keith - livesey - cobb | 64 | 33_morality_moral_keith_livesey | | 34 | nmm - traffic - behind - bike - lane | 61 | 34_nmm_traffic_behind_bike | | 35 | games - sega - genesis - snes - cd | 61 | 35_games_sega_genesis_snes | | 36 | swap - memory - emm386 - windows - file | 61 | 36_swap_memory_emm386_windows | | 37 | president - stephanopoulos - myers - mr - ms | 60 | 37_president_stephanopoulos_myers_mr | | 38 | mary - she - her - immaculate - sin | 60 | 38_mary_she_her_immaculate | | 39 | hst - mission - servicing - solar - shuttle | 59 | 39_hst_mission_servicing_solar | | 40 | copy - protected - protection - disks - sehari | 59 | 40_copy_protected_protection_disks | | 41 | bmw - moa - rider - cactus - requests | 58 | 41_bmw_moa_rider_cactus | | 42 | colormap - dpy - visual - color - window | 58 | 42_colormap_dpy_visual_color | | 43 | points - sphere - den - p3 - p1 | 57 | 43_points_sphere_den_p3 | | 44 | batf - warrant - assault - waco - they | 56 | 44_batf_warrant_assault_waco | | 45 | nsa - encryption - cryptosystems - sternlight - government | 56 | 45_nsa_encryption_cryptosystems_sternlight | | 46 | israel - lebanese - lebanon - israeli - hernlem | 55 | 46_israel_lebanese_lebanon_israeli | | 47 | gaza - israel - palestinian - israeli - peace | 55 | 47_gaza_israel_palestinian_israeli | | 48 | yuk - motorcycling - east - rtsg - riders | 55 | 48_yuk_motorcycling_east_rtsg | | 49 | science - methodology - scientific - sas - fulk | 54 | 49_science_methodology_scientific_sas | | 50 | shift - shifting - manual - transmission - automatic | 53 | 50_shift_shifting_manual_transmission | | 51 | tax - taxes - income - vat - deficit | 53 | 51_tax_taxes_income_vat | | 52 | window - manager - main_win - xsizehints - expose | 52 | 52_window_manager_main_win_xsizehints | | 53 | drive - controller - drives - disk - ide | 52 | 53_drive_controller_drives_disk | | 54 | gif - format - linux - convert - files | 52 | 54_gif_format_linux_convert | | 55 | israeli - israel - hamid - mcrcim - israelis | 51 | 55_israeli_israel_hamid_mcrcim | | 56 | pin - ethernet - board - card - asante | 51 | 56_pin_ethernet_board_card | | 57 | gamma - bursters - oort - ray - cloud | 51 | 57_gamma_bursters_oort_ray | | 58 | drive - floptical - drives - disks - hard | 50 | 58_drive_floptical_drives_disks | | 59 | serial - modem - dtr - uart - rts | 50 | 59_serial_modem_dtr_uart | | 60 | finland - sweden - ericsson - czech - finnish | 49 | 60_finland_sweden_ericsson_czech | | 61 | lankford - torre - he - gilkey - hitter | 49 | 61_lankford_torre_he_gilkey | | 62 | cd - rom - toshiba - cd300 - cdrom | 47 | 62_cd_rom_toshiba_cd300 | | 63 | dog - dogs - springer - dod - bike | 47 | 63_dog_dogs_springer_dod | | 64 | clutch - runs - hit - batting - rbis | 47 | 64_clutch_runs_hit_batting | | 65 | candida - yeast - noring - systemic - infections | 47 | 65_candida_yeast_noring_systemic | | 66 | lopez - year - he - catchers - players | 46 | 66_lopez_year_he_catchers | | 67 | battery - batteries - concrete - acid - lead | 46 | 67_battery_batteries_concrete_acid | | 68 | 50 - 486 - 486dx2 - cyrix - mhz | 46 | 68_50_486_486dx2_cyrix | | 69 | scsi - ide - dma - bus - controller | 46 | 69_scsi_ide_dma_bus | | 70 | font - fonts - truetype - atm - tt | 45 | 70_font_fonts_truetype_atm | | 71 | drugs - drug - cocaine - illegal - marijuana | 45 | 71_drugs_drug_cocaine_illegal | | 72 | helmet - helmets - shoei - jacket - fit | 44 | 72_helmet_helmets_shoei_jacket | | 73 | mormon - mormons - lds - church - ceremonies | 44 | 73_mormon_mormons_lds_church | | 74 | br - isc - steveh - thor - government | 44 | 74_br_isc_steveh_thor | | 75 | allergy - antihistamine - shots - dyer - sleep | 44 | 75_allergy_antihistamine_shots_dyer | | 76 | pens - caps - cup - jets - canucks | 44 | 76_pens_caps_cup_jets | | 77 | petch - god - love - gvg47 - gvg | 44 | 77_petch_god_love_gvg47 | | 78 | mazda - toyota - miles - car - camry | 44 | 78_mazda_toyota_miles_car | | 79 | truth - arrogance - absolutes - absolute - christians | 43 | 79_truth_arrogance_absolutes_absolute | | 80 | shaft - wheelies - stafford - wheelie - winona | 43 | 80_shaft_wheelies_stafford_wheelie | | 81 | crypt - key - cryptography - des - ciphers | 43 | 81_crypt_key_cryptography_des | | 82 | oil - drain - changing - ohio - plug | 42 | 82_oil_drain_changing_ohio | | 83 | jewish - baseball - vb30 - lafibm - players | 42 | 83_jewish_baseball_vb30_lafibm | | 84 | sleeve - sale - picture - cd - 45 | 42 | 84_sleeve_sale_picture_cd | | 85 | morris - team - jays - maynard - viola | 42 | 85_morris_team_jays_maynard | | 86 | cable - antenna - receiver - distance - tv | 41 | 86_cable_antenna_receiver_distance | | 87 | black - king - kyle - adjective - kkopp | 41 | 87_black_king_kyle_adjective | | 88 | countersteering - mjs - bike - countersteering_faq - lean | 41 | 88_countersteering_mjs_bike_countersteering_faq | | 89 | cpu - fan - heat - sink - fans | 41 | 89_cpu_fan_heat_sink | | 90 | jesus - tomb - magi - resurrection - disciples | 41 | 90_jesus_tomb_magi_resurrection | | 91 | canon - scripture - books - bible - septuagint | 40 | 91_canon_scripture_books_bible | | 92 | mac - disks - 800k - 44mb - read | 40 | 92_mac_disks_800k_44mb | | 93 | keenan - rangers - hockey - messier - roger | 40 | 93_keenan_rangers_hockey_messier | | 94 | xv - bit - 24bit - image - images | 39 | 94_xv_bit_24bit_image | | 95 | greek - greece - turkish - greeks - turks | 39 | 95_greek_greece_turkish_greeks | | 96 | drive - meg - ram - sale - scherf | 39 | 96_drive_meg_ram_sale | | 97 | photography - krillean - kirlian - pictures - unlv | 39 | 97_photography_krillean_kirlian_pictures | | 98 | monitors - hours - day - nevai - monitor | 39 | 98_monitors_hours_day_nevai | | 99 | card - orchid - p9000 - vlb - cards | 38 | 99_card_orchid_p9000_vlb | | 100 | sale - list - 00 - guide - shipping | 38 | 100_sale_list_00_guide | | 101 | monitor - screen - problem - 610 - video | 38 | 101_monitor_screen_problem_610 | | 102 | baptism - sin - aaron - infants - baptized | 38 | 102_baptism_sin_aaron_infants | | 103 | kuwait - saudi - iraq - gulf - war | 37 | 103_kuwait_saudi_iraq_gulf | | 104 | station - redesign - dc - shuttle - space | 37 | 104_station_redesign_dc_shuttle | | 105 | marriage - married - marry - ceremony - marriages | 37 | 105_marriage_married_marry_ceremony | | 106 | polygon - polygons - ___ - routine - fast | 37 | 106_polygon_polygons_____routine | | 107 | space - shuttle - launch - afit - astronomy | 37 | 107_space_shuttle_launch_afit | | 108 | sabres - buffalo - fuhr - boston - bruins | 36 | 108_sabres_buffalo_fuhr_boston | | 109 | waco - reno - federal - fbi - batf | 36 | 109_waco_reno_federal_fbi | | 110 | bike - 805 - motorcycle - ride - motorcycles | 36 | 110_bike_805_motorcycle_ride | | 111 | phone - hook - number - line - tip | 36 | 111_phone_hook_number_line | | 112 | phillies - phils - 1964 - bunning - reds | 36 | 112_phillies_phils_1964_bunning | | 113 | roby - fbi - udel - chopin - compound | 35 | 113_roby_fbi_udel_chopin | | 114 | hernia - pain - bone - radiologist - arm | 35 | 114_hernia_pain_bone_radiologist | | 115 | sco - split - newsgroup - graphics - comp | 35 | 115_sco_split_newsgroup_graphics | | 116 | irq - interrupt - port - com4 - com3 | 34 | 116_irq_interrupt_port_com4 | | 117 | gopher - search - images - ftp - data | 34 | 117_gopher_search_images_ftp | | 118 | 3d - grafsys - library - graphics - shading | 34 | 118_3d_grafsys_library_graphics | | 119 | comet - jupiter - gehrels - orbit - sq | 34 | 119_comet_jupiter_gehrels_orbit | | 120 | gtoal - celp - speech - compression - voice | 34 | 120_gtoal_celp_speech_compression | | 121 | insurance - health - private - care - gld | 34 | 121_insurance_health_private_care | | 122 | centaur - proton - energy - uranium - ryukoku | 34 | 122_centaur_proton_energy_uranium | | 123 | easter - goddess - mithras - resurrection - pagan | 33 | 123_easter_goddess_mithras_resurrection | | 124 | cult - cults - freemasonry - baptists - baptist | 32 | 124_cult_cults_freemasonry_baptists | | 125 | ticket - airline - hotel - tickets - voucher | 32 | 125_ticket_airline_hotel_tickets | | 126 | nhl - stars - team - minnesota - franchise | 32 | 126_nhl_stars_team_minnesota | | 127 | sox - red - bosio - bosox - clemens | 32 | 127_sox_red_bosio_bosox | | 128 | ashok - slip - packet - cwru - slipper | 32 | 128_ashok_slip_packet_cwru | | 129 | jehovah - elohim - lord - pope - father | 32 | 129_jehovah_elohim_lord_pope | | 130 | spacecraft - baalke - mission - galileo - pluto | 31 | 130_spacecraft_baalke_mission_galileo | | 131 | speed - 680x0 - x86 - clock - 68040 | 31 | 131_speed_680x0_x86_clock | | 132 | escrow - key - agencies - keys - secure | 31 | 132_escrow_key_agencies_keys | | 133 | doctor - clinic - surgery - patient - japanese | 31 | 133_doctor_clinic_surgery_patient | | 134 | bike - bikes - mower - sale - honda | 31 | 134_bike_bikes_mower_sale | | 135 | wave - bikers - cage - squid - waved | 31 | 135_wave_bikers_cage_squid | | 136 | insurance - fault - car - hail - rates | 31 | 136_insurance_fault_car_hail | | 137 | garrett - ingres - ibm - rickert - turkey | 30 | 137_garrett_ingres_ibm_rickert | | 138 | theism - fanatism - frank - dwyer - belief | 30 | 138_theism_fanatism_frank_dwyer | | 139 | migraine - pain - migraines - zisfein - headache | 30 | 139_migraine_pain_migraines_zisfein | | 140 | 130 - boyle - road - speed - roads | 28 | 140_130_boyle_road_speed | | 141 | satellite - digex - satellites - access - drag | 28 | 141_satellite_digex_satellites_access | | 142 | 610 - centris - iivx - lciii - c610 | 28 | 142_610_centris_iivx_lciii | | 143 | depression - prozac - thyroid - thyroxin - nutrition | 28 | 143_depression_prozac_thyroid_thyroxin | | 144 | journalism - baseball - dwarner - bolick - dodgers | 28 | 144_journalism_baseball_dwarner_bolick | | 145 | tempest - holland - northeastern - monitor - colostate | 28 | 145_tempest_holland_northeastern_monitor | | 146 | 00 - wolverine - 1st - 50 - comics | 28 | 146_00_wolverine_1st_50 | | 147 | murray - gm - wings - ottawa - lindros | 28 | 147_murray_gm_wings_ottawa | | 148 | duo - 230 - beeps - chimes - machine | 27 | 148_duo_230_beeps_chimes | | 149 | mr2 - engine - clutch - eliot - noisy | 27 | 149_mr2_engine_clutch_eliot | | 150 | christianity - convenient - christian - definition - christians | 27 | 150_christianity_convenient_christian_definition | | 151 | satan - ra - god - lucifer - heaven | 27 | 151_satan_ra_god_lucifer | | 152 | summer - room - sublet - jhuvm - bedroom | 26 | 152_summer_room_sublet_jhuvm | | 153 | software - wingert - level - sci - space | 26 | 153_software_wingert_level_sci | | 154 | god - jesus - malcolm - royalroads - law | 26 | 154_god_jesus_malcolm_royalroads | | 155 | europeans - nhl - rauser - players - european | 26 | 155_europeans_nhl_rauser_players | | 156 | mustang - camaro - ford - howell - firebird | 25 | 156_mustang_camaro_ford_howell | | 157 | stove - wpi - irvine - stratus - electric | 25 | 157_stove_wpi_irvine_stratus | | 158 | scope - scopes - oscilloscope - fluke - phosphor | 25 | 158_scope_scopes_oscilloscope_fluke | | 159 | odometer - bmw - sensor - car - dealer | 25 | 159_odometer_bmw_sensor_car | | 160 | koresh - utarlg - sbc - uta - backing | 25 | 160_koresh_utarlg_sbc_uta | | 161 | tape - backup - adaptec - aspi4dos - 1542 | 25 | 161_tape_backup_adaptec_aspi4dos | | 162 | mask - goalie - gtd597a - votes - hrivnak | 25 | 162_mask_goalie_gtd597a_votes | | 163 | astros - houston - games - rbi - sweda | 24 | 163_astros_houston_games_rbi | | 164 | icon - icons - program - manager - vpnet | 24 | 164_icon_icons_program_manager | | 165 | solvent - adhesive - duct - tape - ruck | 24 | 165_solvent_adhesive_duct_tape | | 166 | keymap - key - numlock - keyboard - xterm | 24 | 166_keymap_key_numlock_keyboard | | 167 | ir - dres - dnd - detector - cycle | 24 | 167_ir_dres_dnd_detector | | 168 | car - dealer - price - blue - sales | 24 | 168_car_dealer_price_blue | | 169 | midi - sound - blaster - driver - soundblaster | 24 | 169_midi_sound_blaster_driver | | 170 | blue - boards - leds - led - green | 24 | 170_blue_boards_leds_led | | 171 | wax - scratches - plastic - finish - paint | 24 | 171_wax_scratches_plastic_finish | | 172 | motif - linux - bindings - xact - cose | 24 | 172_motif_linux_bindings_xact | | 173 | v4 - v12 - cdac - v8 - ole | 24 | 173_v4_v12_cdac_v8 | | 174 | officers - cop - mcguire - xxxx - police | 23 | 174_officers_cop_mcguire_xxxx | | 175 | gant - hirschbeck - umpire - strike - duke | 23 | 175_gant_hirschbeck_umpire_strike | | 176 | abortion - abortions - nyikos - choice - landreneau | 23 | 176_abortion_abortions_nyikos_choice | | 177 | sharks - season - chuq - grade - acquired | 23 | 177_sharks_season_chuq_grade | | 178 | punishment - penalty - capital - death - innocent | 23 | 178_punishment_penalty_capital_death | | 179 | mouse - windows - driver - stuttgart - com3 | 23 | 179_mouse_windows_driver_stuttgart | | 180 | processing - image - imaging - mishra - hendrix | 23 | 180_processing_image_imaging_mishra | | 181 | freedom - virginia - beyer - ucla - ab4z | 23 | 181_freedom_virginia_beyer_ucla | | 182 | seizures - corn - paulson - seizure - cereals | 23 | 182_seizures_corn_paulson_seizure | | 183 | crohn - ibd - inflammation - diet - wiesel | 23 | 183_crohn_ibd_inflammation_diet | | 184 | barbecued - foods - carcinogenic - food - meat | 23 | 184_barbecued_foods_carcinogenic_food | | 185 | pillion - riding - advice - passenger - ride | 22 | 185_pillion_riding_advice_passenger | | 186 | key - chip - clipper - session - encrypted | 22 | 186_key_chip_clipper_session | | 187 | powerbook - portable - pb100 - pb - peirce | 22 | 187_powerbook_portable_pb100_pb | | 188 | ear - ears - hearing - earwax - dizziness | 22 | 188_ear_ears_hearing_earwax | | 189 | photoshop - adobe - rot - dgf1 - qc | 22 | 189_photoshop_adobe_rot_dgf1 | | 190 | evolution - theory - rawlins - scharle - science | 22 | 190_evolution_theory_rawlins_scharle | | 191 | ftp - nonibm - puff - glp - minivas | 22 | 191_ftp_nonibm_puff_glp | | 192 | scanner - scanners - logitech - scanman - grayscale | 22 | 192_scanner_scanners_logitech_scanman | | 193 | games - baseball - game - pitches - pitcher | 22 | 193_games_baseball_game_pitches | | 194 | ham - interference - surges - alternator - watts | 22 | 194_ham_interference_surges_alternator | | 195 | weight - omen - chromium - diet - fat | 22 | 195_weight_omen_chromium_diet | | 196 | pregnency - teacher - oswego - biology - sperm | 21 | 196_pregnency_teacher_oswego_biology | | 197 | ghostscript - postscript - ghostview - pageview - ftms | 21 | 197_ghostscript_postscript_ghostview_pageview | | 198 | 3do - 3d - lightwave - list - imagine | 21 | 198_3do_3d_lightwave_list | | 199 | polio - disease - alzheimer - syndrome - patients | 21 | 199_polio_disease_alzheimer_syndrome | | 200 | motherboard - 386 - 386dx - murli - sale | 21 | 200_motherboard_386_386dx_murli | | 201 | des - key - bits - block - attack | 21 | 201_des_key_bits_block | | 202 | ax - max - g9v - b8f - a86 | 21 | 202_ax_max_g9v_b8f | | 203 | israeli - biased - israel - media - none | 21 | 203_israeli_biased_israel_media | | 204 | exhaust - carbs - intake - engine - air | 21 | 204_exhaust_carbs_intake_engine | | 205 | tickets - 05pm - 35pm - june - ticket | 21 | 205_tickets_05pm_35pm_june | | 206 | chain - wax - maxima - cookson - mitre | 21 | 206_chain_wax_maxima_cookson | | 207 | toyota - cruiser - suv - jeep - explorer | 21 | 207_toyota_cruiser_suv_jeep | | 208 | lipman - visualization - navy - graphics - seminar | 20 | 208_lipman_visualization_navy_graphics | | 209 | dwi - speedy - driving - svoboda - liquor | 20 | 209_dwi_speedy_driving_svoboda | | 210 | dialing - phones - tone - hugo - sweden | 20 | 210_dialing_phones_tone_hugo | | 211 | image - processing - plplot - analysis - plotting | 20 | 211_image_processing_plplot_analysis | | 212 | convertible - wife - car - targa - convertibles | 20 | 212_convertible_wife_car_targa | | 213 | vuille - babb - synapse - ic - pcmcia | 20 | 213_vuille_babb_synapse_ic | | 214 | nt - windows - chicogo - os - reimert | 20 | 214_nt_windows_chicogo_os | | 215 | alomar - defensive - sandberg - average - career | 20 | 215_alomar_defensive_sandberg_average | | 216 | blues - hawks - joseph - blackhawks - shanahan | 20 | 216_blues_hawks_joseph_blackhawks | | 217 | graphics - pub - 128 - ray - ftp | 20 | 217_graphics_pub_128_ray | | 218 | w4wg - network - windows - workgroups - lastdrive | 20 | 218_w4wg_network_windows_workgroups | | 219 | tank - bag - goldberg - fj1100 - pouch | 20 | 219_tank_bag_goldberg_fj1100 | | 220 | mailing - list - detweiler - mail - rdetweil | 20 | 220_mailing_list_detweiler_mail | | 221 | gas - tear - unb - riddle - j979 | 20 | 221_gas_tear_unb_riddle | | 222 | ide - bus - controller - vlb - scsi | 20 | 222_ide_bus_controller_vlb | | 223 | saturn - dealer - profit - warranty - sl2 | 19 | 223_saturn_dealer_profit_warranty | | 224 | cursor - xterm - blinking - cursors - allbery | 19 | 224_cursor_xterm_blinking_cursors | | 225 | joystick - joysticks - arcade - port - int15h | 19 | 225_joystick_joysticks_arcade_port | | 226 | lyme - disease - ld - infectious - patients | 19 | 226_lyme_disease_ld_infectious | | 227 | context - jim - joslin - meritt - mwunix | 19 | 227_context_jim_joslin_meritt | | 228 | qualcomm - clinton - qualcom - rdippold - clipper | 19 | 228_qualcomm_clinton_qualcom_rdippold | | 229 | cancer - hiv - burzynski - breast - booklet | 19 | 229_cancer_hiv_burzynski_breast | | 230 | kidney - stones - calcium - she - stone | 19 | 230_kidney_stones_calcium_she | | 231 | rosicrucian - amorc - ch981 - alicea - tony | 19 | 231_rosicrucian_amorc_ch981_alicea | | 232 | henrik - armenia - bm - armenians - karabakh | 19 | 232_henrik_armenia_bm_armenians | | 233 | geico - insurance - claim - davew - wonnacott | 19 | 233_geico_insurance_claim_davew | | 234 | eye - dominance - prk - handedness - rk | 19 | 234_eye_dominance_prk_handedness | | 235 | church - churches - crossroads - movement - boston | 19 | 235_church_churches_crossroads_movement | | 236 | water - mwra - phd - cellar - scoggin | 19 | 236_water_mwra_phd_cellar | | 237 | integra - car - shadow - dodge - gtz | 19 | 237_integra_car_shadow_dodge | | 238 | sabbath - worship - law - ceremonial - paul | 19 | 238_sabbath_worship_law_ceremonial | | 239 | lobby - sammons - letter - ns111310 - colostate | 19 | 239_lobby_sammons_letter_ns111310 | | 240 | henry - orion - film - prototype - toronto | 18 | 240_henry_orion_film_prototype | | 241 | trinity - father - son - holy - god | 18 | 241_trinity_father_son_holy | | 242 | captain - traded - captains - striped - resigned | 18 | 242_captain_traded_captains_striped | | 243 | 42 - tiff - philosophical - significance - joachim | 18 | 243_42_tiff_philosophical_significance | | 244 | space - mars - spaceflight - nick - fred | 18 | 244_space_mars_spaceflight_nick | | 245 | astronaut - space - nasa - pilot - jemison | 18 | 245_astronaut_space_nasa_pilot | | 246 | circumcision - cons - pros - penile - blix | 18 | 246_circumcision_cons_pros_penile | | 247 | wire - wiring - ground - neutral - outlets | 17 | 247_wire_wiring_ground_neutral | | 248 | women - men - monash - depression - sex | 17 | 248_women_men_monash_depression | | 249 | prophecy - prophecies - earthquake - lord - prophesies | 17 | 249_prophecy_prophecies_earthquake_lord | | 250 | cooling - towers - nuclear - plants - water | 17 | 250_cooling_towers_nuclear_plants | | 251 | diesel - diesels - fuel - injector - emissions | 17 | 251_diesel_diesels_fuel_injector | | 252 | windows - pif - dos - file - command | 17 | 252_windows_pif_dos_file | | 253 | uv - bulb - flashlight - bulbs - neon | 17 | 253_uv_bulb_flashlight_bulbs | | 254 | tires - tire - fluids - abs - dot | 17 | 254_tires_tire_fluids_abs | | 255 | mhz - clock - operational - iisi - cpu | 17 | 255_mhz_clock_operational_iisi | | 256 | cubs - braves - team - america - talent | 17 | 256_cubs_braves_team_america | | 257 | lens - rupin - camera - dang - dartmouth | 17 | 257_lens_rupin_camera_dang | | 258 | dock - duo - apple - bredell - deguzman | 16 | 258_dock_duo_apple_bredell | | 259 | janet - reno - madman - children - she | 16 | 259_janet_reno_madman_children | | 260 | lock - locks - cobra - kryptonite - cable | 16 | 260_lock_locks_cobra_kryptonite | | 261 | mouse - jumpy - motion - byu - smoothly | 16 | 261_mouse_jumpy_motion_byu | | 262 | god - creates - omnipotence - shaped - omnipotent | 16 | 262_god_creates_omnipotence_shaped | | 263 | yassin - deir - irgun - dir - village | 16 | 263_yassin_deir_irgun_dir | | 264 | xv - julian - copyright - lancs - escaped | 16 | 264_xv_julian_copyright_lancs | | 265 | mjm - fm - circuits - mixer - musone | 16 | 265_mjm_fm_circuits_mixer | | 266 | tga - rle - pov - povray - tmp | 16 | 266_tga_rle_pov_povray | | 267 | workspace - managers - workspaces - manager - zip | 16 | 267_workspace_managers_workspaces_manager | | 268 | quadra - scsi - nodine - cartridge - mac | 16 | 268_quadra_scsi_nodine_cartridge | | 269 | hpgl - ilmenau - naplps - vuw - schmidt | 16 | 269_hpgl_ilmenau_naplps_vuw | | 270 | jumper - 2190 - maxtor - thad - drive | 16 | 270_jumper_2190_maxtor_thad | | 271 | dxf - iff - format - autocad - pei | 16 | 271_dxf_iff_format_autocad | | 272 | mode - vesa - vga - svga - 640x400 | 16 | 272_mode_vesa_vga_svga | | 273 | mosques - mosque - jerusalem - eggertj - jake | 16 | 273_mosques_mosque_jerusalem_eggertj | | 274 | ulf - erau - huot - players - drozinst | 15 | 274_ulf_erau_huot_players | | 275 | algorithm - secret - chip - reverse - clipper | 15 | 275_algorithm_secret_chip_reverse | | 276 | font - fonts - alavi - ssa - 8514 | 15 | 276_font_fonts_alavi_ssa | | 277 | gauge - nancy - gauges - temp - cigarette | 15 | 277_gauge_nancy_gauges_temp | | 278 | octopus - detroit - ice - cunyvm - hammerl | 15 | 278_octopus_detroit_ice_cunyvm | | 279 | cview - temp - moscom - directory - zenkar | 15 | 279_cview_temp_moscom_directory | | 280 | drive - cable - quantum - disk - internal | 15 | 280_drive_cable_quantum_disk | | 281 | logo - vgalogo - rle - startup - lgo | 15 | 281_logo_vgalogo_rle_startup | | 282 | ini - updating - svein - sysedit - utility | 15 | 282_ini_updating_svein_sysedit | | 283 | sin - hate - sinner - love - scott | 15 | 283_sin_hate_sinner_love | | 284 | administration - privacy - eff - government - inquiry | 15 | 284_administration_privacy_eff_government | | 285 | bonds - williams - batting - giants - clark | 15 | 285_bonds_williams_batting_giants | | 286 | 02106 - chemistry - udel - paperback - ravel | 15 | 286_02106_chemistry_udel_paperback | | 287 | cherry - coach - don - he - him | 15 | 287_cherry_coach_don_he | | 288 | drink - drinking - riding - alcohol - hours | 15 | 288_drink_drinking_riding_alcohol | | 289 | ether - planets - twist - sci - mnemonics | 14 | 289_ether_planets_twist_sci | | 290 | keys - des - lokkur - nanosecond - keyseach | 14 | 290_keys_des_lokkur_nanosecond | | 291 | virginia - uva - partying - andi - schools | 14 | 291_virginia_uva_partying_andi | | 292 | hiram - dk - vhs - kou - koutd | 14 | 292_hiram_dk_vhs_kou | | 293 | eliot - flat - boxer - 180 - v12 | 14 | 293_eliot_flat_boxer_180 | | 294 | neilson - triumf - seoul - deadly - kids | 14 | 294_neilson_triumf_seoul_deadly | | 295 | cruel - keith - caltech - constitution - painful | 14 | 295_cruel_keith_caltech_constitution | | 296 | luminosity - red - rgb - hue - green | 14 | 296_luminosity_red_rgb_hue | | 297 | she - were - they - her - sumgait | 14 | 297_she_were_they_her | | 298 | jagr - francis - minus - uvic - player | 14 | 298_jagr_francis_minus_uvic | | 299 | adl - bullock - gerard - francisco - arens | 14 | 299_adl_bullock_gerard_francisco | | 300 | widgets - gadgets - dealy - motif - widget | 14 | 300_widgets_gadgets_dealy_motif | | 301 | print - printer - file - claebaur - portal | 14 | 301_print_printer_file_claebaur | | 302 | hacker - ethic - hackers - dorsai - carlos | 14 | 302_hacker_ethic_hackers_dorsai | | 303 | weick - dana - him - cpu - sturges | 14 | 303_weick_dana_him_cpu | | 304 | xputimage - server - sunview - cam - animation | 14 | 304_xputimage_server_sunview_cam | | 305 | god - evil - serbian - saved - genocide | 14 | 305_god_evil_serbian_saved | | 306 | nubus - pds - lc - marvin - higgins | 13 | 306_nubus_pds_lc_marvin | | 307 | zeos - gateway - murthy - service - vasudev | 13 | 307_zeos_gateway_murthy_service | | 308 | temperature - henry - interstellar - sky - radiation | 13 | 308_temperature_henry_interstellar_sky | | 309 | uniforms - marlins - lloyd - reiniger - reds | 13 | 309_uniforms_marlins_lloyd_reiniger | | 310 | faith - saved - romans - lukewarm - deeds | 13 | 310_faith_saved_romans_lukewarm | | 311 | scsi - drive - ide - oracle - adaptec | 13 | 311_scsi_drive_ide_oracle | | 312 | fifth - keyphrase - amendment - key - passwords | 13 | 312_fifth_keyphrase_amendment_key | | 313 | tongues - language - tounges - gifted - bjorn | 13 | 313_tongues_language_tounges_gifted | | 314 | rocks - overpass - ejv2j - erik - kids | 13 | 314_rocks_overpass_ejv2j_erik | | 315 | biggest - disappointment - smale - mvp - surprise | 13 | 315_biggest_disappointment_smale_mvp | | 316 | nicknames - nickname - healy - tammy - berg | 13 | 316_nicknames_nickname_healy_tammy | | 317 | ampere - amp - db - bell - ohmite | 13 | 317_ampere_amp_db_bell | | 318 | handling - ntuvax - ntu - handson - ba7116326 | 13 | 318_handling_ntuvax_ntu_handson | | 319 | air - r12 - conditioning - substitutes - freon | 13 | 319_air_r12_conditioning_substitutes | | 320 | soenke - bielefeld - widget - savela - masc0442 | 13 | 320_soenke_bielefeld_widget_savela | | 321 | eisa - isa - bus - board - video | 13 | 321_eisa_isa_bus_board | | 322 | wrench - srb - thiokol - pliers - tool | 13 | 322_wrench_srb_thiokol_pliers | | 323 | oilers - pocklington - edmonton - northlands - yadallee | 13 | 323_oilers_pocklington_edmonton_northlands | | 324 | sound - stereo - channel - mac - soundbase | 13 | 324_sound_stereo_channel_mac | | 325 | movies - bikes - csundh30 - cassidy - ursa | 13 | 325_movies_bikes_csundh30_cassidy | | 326 | haldol - elderly - lithium - drugs - hospital | 13 | 326_haldol_elderly_lithium_drugs | | 327 | 8051 - oscar - mont - speth - spock | 13 | 327_8051_oscar_mont_speth | | 328 | cache - iisi - powercache - card - fpu | 13 | 328_cache_iisi_powercache_card | | 329 | bryce - bike - manish - arches - touring | 13 | 329_bryce_bike_manish_arches | | 330 | skate - carol - malarchuk - sei - neck | 13 | 330_skate_carol_malarchuk_sei | | 331 | rush - compuserve - jongsma - anovak - henson | 12 | 331_rush_compuserve_jongsma_anovak | | 332 | date - clock - dos - bios - cmos | 12 | 332_date_clock_dos_bios | | 333 | mcadams - sale - suresh - mattress - aj008 | 12 | 333_mcadams_sale_suresh_mattress | | 334 | silence - moment - prayer - eeb1 - opposing | 12 | 334_silence_moment_prayer_eeb1 | | 335 | jesus - prayers - god - name - prayer | 12 | 335_jesus_prayers_god_name | | 336 | habitable - planets - atmosphere - oxygen - planet | 12 | 336_habitable_planets_atmosphere_oxygen | | 337 | sunset - sunrise - drexel - cbis - wetstein | 12 | 337_sunset_sunrise_drexel_cbis | | 338 | selective - borden - service - abolish - naval | 12 | 338_selective_borden_service_abolish | | 339 | illustrator - diablo - autotrace - points - drawing | 12 | 339_illustrator_diablo_autotrace_points | | 340 | love - kodak - god - dps - ico | 12 | 340_love_kodak_god_dps | | 341 | koresh - griffen - batf - children - fbi | 12 | 341_koresh_griffen_batf_children | | 342 | needles - acupuncture - needle - aids - hypodermic | 12 | 342_needles_acupuncture_needle_aids | | 343 | accelerations - acceleration - 45g - deaddio - amruth | 12 | 343_accelerations_acceleration_45g_deaddio | | 344 | tape - copy - vcr - video - destructing | 12 | 344_tape_copy_vcr_video | | 345 | pmy - sword - royalroads - yadlowsky - malcolm | 12 | 345_pmy_sword_royalroads_yadlowsky | | 346 | educational - price - newsbytes - cda - eu | 12 | 346_educational_price_newsbytes_cda | | 347 | liar - lunatic - he - christian - bible | 11 | 347_liar_lunatic_he_christian | | 348 | eff - minerva - yale - jgfoot - tarl | 11 | 348_eff_minerva_yale_jgfoot | | 349 | seema - hannover - madvlsi - varma - columbia | 11 | 349_seema_hannover_madvlsi_varma | | 350 | eugenics - memes - genes - genome - ruegg | 11 | 350_eugenics_memes_genes_genome | | 351 | lunar - ltm1 - manned - tele - exploration | 11 | 351_lunar_ltm1_manned_tele | | 352 | switch - beams - st11 - bimmer - cookson | 11 | 352_switch_beams_st11_bimmer | | 353 | commandment - christians - temper - inference - jesus | 11 | 353_commandment_christians_temper_inference | | 354 | harkey - dl - oscs - cubs - wetteland | 11 | 354_harkey_dl_oscs_cubs | | 355 | fourd - 0565 - 494 - dimension - cute | 11 | 355_fourd_0565_494_dimension | | 356 | mattingly - tesla - njit - drm6640 - baseman | 11 | 356_mattingly_tesla_njit_drm6640 | | 357 | placebo - roth - rr - medicine - jb | 11 | 357_placebo_roth_rr_medicine | | 358 | tempest - c650 - cyclone - price - drop | 10 | 358_tempest_c650_cyclone_price | | 359 | ssf - overhead - nasa - tax - billion | 10 | 359_ssf_overhead_nasa_tax | | 360 | mining - freaks - alaska - eco - miners | 10 | 360_mining_freaks_alaska_eco | </details> ## Training hyperparameters * calculate_probabilities: False * language: None * low_memory: False * min_topic_size: 10 * n_gram_range: (1, 1) * nr_topics: None * seed_topic_list: None * top_n_words: 10 * verbose: False * zeroshot_min_similarity: 0.7 * zeroshot_topic_list: None ## Framework versions * Numpy: 1.25.2 * HDBSCAN: 0.8.33 * UMAP: 0.5.6 * Pandas: 2.0.3 * Scikit-Learn: 1.2.2 * Sentence-transformers: 2.7.0 * Transformers: 4.40.1 * Numba: 0.58.1 * Plotly: 5.15.0 * Python: 3.10.12
farrukhrasool112/distillbert-finetuned-ner-btc
farrukhrasool112
2024-05-08T08:46:38Z
105
0
transformers
[ "transformers", "safetensors", "distilbert", "token-classification", "generated_from_trainer", "base_model:distilbert/distilbert-base-cased", "base_model:finetune:distilbert/distilbert-base-cased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2024-03-12T09:37:46Z
--- license: apache-2.0 base_model: distilbert/distilbert-base-cased tags: - generated_from_trainer metrics: - precision - recall - f1 - accuracy model-index: - name: distillbert-finetuned-ner-btc results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distillbert-finetuned-ner-btc This model is a fine-tuned version of [distilbert/distilbert-base-cased](https://huggingface.co/distilbert/distilbert-base-cased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.2802 - Precision: 0.4392 - Recall: 0.4370 - F1: 0.4381 - Accuracy: 0.9117 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.0 | 56 | 0.3448 | 0.4762 | 0.0345 | 0.0644 | 0.8792 | | No log | 2.0 | 112 | 0.2880 | 0.4506 | 0.3938 | 0.4203 | 0.9080 | | No log | 3.0 | 168 | 0.2802 | 0.4392 | 0.4370 | 0.4381 | 0.9117 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.0.0+cu117 - Datasets 2.18.0 - Tokenizers 0.15.2
annamalai-s/bertopic_newsgroup_mpnet
annamalai-s
2024-05-08T08:46:04Z
8
0
bertopic
[ "bertopic", "text-classification", "region:us" ]
text-classification
2024-05-08T08:46:02Z
--- tags: - bertopic library_name: bertopic pipeline_tag: text-classification --- # bertopic_newsgroup_mpnet This is a [BERTopic](https://github.com/MaartenGr/BERTopic) model. BERTopic is a flexible and modular topic modeling framework that allows for the generation of easily interpretable topics from large datasets. ## Usage To use this model, please install BERTopic: ``` pip install -U bertopic ``` You can use the model as follows: ```python from bertopic import BERTopic topic_model = BERTopic.load("annamalai-s/bertopic_newsgroup_mpnet") topic_model.get_topic_info() ``` ## Topic overview * Number of topics: 445 * Number of training documents: 18846 <details> <summary>Click here for an overview of all topics.</summary> | Topic ID | Topic Keywords | Topic Frequency | Label | |----------|----------------|-----------------|-------| | -1 | to - the - for - from - is | 10 | -1_to_the_for_from | | 0 | gun - guns - firearms - crime - handgun | 5381 | 0_gun_guns_firearms_crime | | 1 | cramer - optilink - gay - clayton - sexual | 266 | 1_cramer_optilink_gay_clayton | | 2 | fbi - batf - waco - compound - koresh | 229 | 2_fbi_batf_waco_compound | | 3 | db - mov - bh - si - bl | 134 | 3_db_mov_bh_si | | 4 | atf - fire - survivors - ranch - dividian | 132 | 4_atf_fire_survivors_ranch | | 5 | drive - slave - drives - master - tape | 127 | 5_drive_slave_drives_master | | 6 | moon - lunar - billion - alaska - prize | 127 | 6_moon_lunar_billion_alaska | | 7 | armenian - turkish - armenians - serdar - argic | 117 | 7_armenian_turkish_armenians_serdar | | 8 | espn - game - abc - coverage - hockey | 112 | 8_espn_game_abc_coverage | | 9 | 3d - phigs - graphics - navy - lipman | 112 | 9_3d_phigs_graphics_navy | | 10 | israeli - israel - israelis - arab - soldiers | 104 | 10_israeli_israel_israelis_arab | | 11 | dos - xfree86 - windows - server - tcp | 100 | 11_dos_xfree86_windows_server | | 12 | sale - drive - meg - ram - floppy | 98 | 12_sale_drive_meg_ram | | 13 | homosexuality - homosexual - paul - christians - sin | 97 | 13_homosexuality_homosexual_paul_christians | | 14 | clutch - alomar - runs - baerga - average | 93 | 14_clutch_alomar_runs_baerga | | 15 | os - microsoft - challenge - supporters - windows | 91 | 15_os_microsoft_challenge_supporters | | 16 | migraine - sleep - dyer - thyroid - geb | 91 | 16_migraine_sleep_dyer_thyroid | | 17 | drive - ide - scsi - drives - controller | 87 | 17_drive_ide_scsi_drives | | 18 | modem - modems - fax - courier - sportster | 85 | 18_modem_modems_fax_courier | | 19 | msg - food - sensitivity - chinese - superstition | 84 | 19_msg_food_sensitivity_chinese | | 20 | objective - morality - larson - frank - values | 84 | 20_objective_morality_larson_frank | | 21 | windows - swap - memory - emm386 - file | 84 | 21_windows_swap_memory_emm386 | | 22 | sale - speakers - stereo - offer - amp | 83 | 22_sale_speakers_stereo_offer | | 23 | jpeg - gif - image - format - file | 82 | 23_jpeg_gif_image_format | | 24 | jewish - zionism - israel - jews - jew | 82 | 24_jewish_zionism_israel_jews | | 25 | space - nasa - venus - planet - earth | 80 | 25_space_nasa_venus_planet | | 26 | encryption - clipper - chip - government - wiretap | 78 | 26_encryption_clipper_chip_government | | 27 | polygon - polygons - ___ - routine - algorithm | 76 | 27_polygon_polygons_____routine | | 28 | car - miles - toyota - sale - mazda | 73 | 28_car_miles_toyota_sale | | 29 | scsi - ide - dma - bus - isa | 71 | 29_scsi_ide_dma_bus | | 30 | 25 - pit - pts - det - la | 69 | 30_25_pit_pts_det | | 31 | stephanopoulos - president - mr - myers - ms | 68 | 31_stephanopoulos_president_mr_myers | | 32 | rushdie - islam - jaeger - islamic - gregg | 66 | 32_rushdie_islam_jaeger_islamic | | 33 | dumbest - automotive - lights - concepts - continental | 66 | 33_dumbest_automotive_lights_concepts | | 34 | motif - openwindows - xview - olit - x11 | 66 | 34_motif_openwindows_xview_olit | | 35 | games - sega - genesis - snes - sale | 65 | 35_games_sega_genesis_snes | | 36 | bosnia - muslims - bosnian - serbs - bosnians | 65 | 36_bosnia_muslims_bosnian_serbs | | 37 | mary - her - she - immaculate - conception | 64 | 37_mary_her_she_immaculate | | 38 | israel - lebanese - lebanon - israeli - hezbollah | 64 | 38_israel_lebanese_lebanon_israeli | | 39 | existence - evolution - theory - science - exist | 62 | 39_existence_evolution_theory_science | | 40 | hell - eternal - heaven - god - jesus | 62 | 40_hell_eternal_heaven_god | | 41 | simms - simm - meg - pin - ram | 61 | 41_simms_simm_meg_pin | | 42 | br - isc - government - steveh - thor | 61 | 42_br_isc_government_steveh | | 43 | dos - stacker - windows - dos6 - disk | 60 | 43_dos_stacker_windows_dos6 | | 44 | clutch - shifting - shift - manual - transmission | 60 | 44_clutch_shifting_shift_manual | | 45 | radar - detector - detectors - valentine - ka | 59 | 45_radar_detector_detectors_valentine | | 46 | tax - taxes - income - deficit - vat | 59 | 46_tax_taxes_income_deficit | | 47 | keyboard - key - keys - keycode - accelerators | 58 | 47_keyboard_key_keys_keycode | | 48 | copy - protected - protection - disks - sehari | 58 | 48_copy_protected_protection_disks | | 49 | station - redesign - space - nasa - option | 57 | 49_station_redesign_space_nasa | | 50 | lib - libxmu - ndet_loop - xmu - usr | 56 | 50_lib_libxmu_ndet_loop_xmu | | 51 | dog - dogs - parr - ucalgary - attack | 55 | 51_dog_dogs_parr_ucalgary | | 52 | leafs - wings - game - detroit - goal | 52 | 52_leafs_wings_game_detroit | | 53 | cancer - water - medical - mwra - circumcision | 51 | 53_cancer_water_medical_mwra | | 54 | sleeve - cd - cds - sale - picture | 51 | 54_sleeve_cd_cds_sale | | 55 | sharks - season - keenan - rangers - chuq | 50 | 55_sharks_season_keenan_rangers | | 56 | battery - batteries - concrete - acid - lead | 49 | 56_battery_batteries_concrete_acid | | 57 | drugs - drug - marijuana - legalization - cigarettes | 49 | 57_drugs_drug_marijuana_legalization | | 58 | exhaust - carbs - bike - carb - honda | 49 | 58_exhaust_carbs_bike_carb | | 59 | shaft - wheelies - stafford - wheelie - winona | 48 | 59_shaft_wheelies_stafford_wheelie | | 60 | key - chip - clipper - algorithm - chips | 48 | 60_key_chip_clipper_algorithm | | 61 | photography - krillean - kirlian - pictures - sol1 | 48 | 61_photography_krillean_kirlian_pictures | | 62 | bike - bikes - motorcycle - motorcycles - buying | 48 | 62_bike_bikes_motorcycle_motorcycles | | 63 | lynn - riders - rtsg - motorcycling - bike | 47 | 63_lynn_riders_rtsg_motorcycling | | 64 | church - churches - christianity - christian - movement | 47 | 64_church_churches_christianity_christian | | 65 | hst - mission - servicing - shuttle - boost | 47 | 65_hst_mission_servicing_shuttle | | 66 | nissan - wagon - villager - altima - vw | 47 | 66_nissan_wagon_villager_altima | | 67 | helmet - helmets - shoei - jacket - eskimo | 47 | 67_helmet_helmets_shoei_jacket | | 68 | finland - sweden - wc - czech - ericsson | 47 | 68_finland_sweden_wc_czech | | 69 | gamma - oort - bursters - ray - cloud | 46 | 69_gamma_oort_bursters_ray | | 70 | jewish - baseball - vb30 - lafibm - players | 46 | 70_jewish_baseball_vb30_lafibm | | 71 | sky - vandalizing - night - pollution - enzo | 46 | 71_sky_vandalizing_night_pollution | | 72 | nanao - monitors - viewsonic - monitor - inches | 46 | 72_nanao_monitors_viewsonic_monitor | | 73 | militia - amendment - arms - regulated - bear | 45 | 73_militia_amendment_arms_regulated | | 74 | rocks - teenagers - warning - overpass - kids | 45 | 74_rocks_teenagers_warning_overpass | | 75 | des - key - keyseach - bits - shelf | 45 | 75_des_key_keyseach_bits | | 76 | dl - wetteland - harkey - franco - plymouth | 45 | 76_dl_wetteland_harkey_franco | | 77 | petch - gvg47 - love - god - gvg | 44 | 77_petch_gvg47_love_god | | 78 | pin - card - connector - ethernet - board | 44 | 78_pin_card_connector_ethernet | | 79 | leds - uv - led - blue - subliminal | 43 | 79_leds_uv_led_blue | | 80 | theism - fanatism - atheism - belief - theists | 42 | 80_theism_fanatism_atheism_belief | | 81 | freedom - forged - locutus - colorado - speech | 42 | 81_freedom_forged_locutus_colorado | | 82 | moral - morality - keith - livesey - caltech | 41 | 82_moral_morality_keith_livesey | | 83 | phd - environmentalism - environmental - heath - pantheism | 41 | 83_phd_environmentalism_environmental_heath | | 84 | buffalo - sabres - blues - bruins - boston | 41 | 84_buffalo_sabres_blues_bruins | | 85 | countersteering - mjs - bike - countersteering_faq - lean | 41 | 85_countersteering_mjs_bike_countersteering_faq | | 86 | nmm - behind - traffic - lane - bike | 41 | 86_nmm_behind_traffic_lane | | 87 | games - game - baseball - pitches - pitcher | 41 | 87_games_game_baseball_pitches | | 88 | cpu - fan - heat - sink - fans | 41 | 88_cpu_fan_heat_sink | | 89 | jehovah - elohim - father - lord - son | 41 | 89_jehovah_elohim_father_lord | | 90 | cruel - punishment - keith - penalty - death | 40 | 90_cruel_punishment_keith_penalty | | 91 | insurance - health - private - care - gld | 40 | 91_insurance_health_private_care | | 92 | powerbook - duo - portable - pb - pb100 | 40 | 92_powerbook_duo_portable_pb | | 93 | bike - sale - miles - mower - fork | 39 | 93_bike_sale_miles_mower | | 94 | postscript - ghostscript - ghostview - pageview - files | 39 | 94_postscript_ghostscript_ghostview_pageview | | 95 | candida - yeast - noring - systemic - infections | 39 | 95_candida_yeast_noring_systemic | | 96 | card - p9000 - orchid - weitek - vlb | 39 | 96_card_p9000_orchid_weitek | | 97 | jews - israel - arabs - land - arab | 38 | 97_jews_israel_arabs_land | | 98 | radiosity - pov - raytracing - ray - amann | 38 | 98_radiosity_pov_raytracing_ray | | 99 | oil - drain - changing - ohio - magnus | 38 | 99_oil_drain_changing_ohio | | 100 | scope - scopes - oscilloscope - fluke - meter | 38 | 100_scope_scopes_oscilloscope_fluke | | 101 | faith - god - exist - proof - burden | 37 | 101_faith_god_exist_proof | | 102 | sox - rbi - games - game - win | 37 | 102_sox_rbi_games_game | | 103 | greek - greece - greeks - turkish - turks | 37 | 103_greek_greece_greeks_turkish | | 104 | science - methodology - sas - fulk - lady | 37 | 104_science_methodology_sas_fulk | | 105 | hockey - nhl - team - league - stars | 37 | 105_hockey_nhl_team_league | | 106 | koresh - fbi - compound - fire - cult | 37 | 106_koresh_fbi_compound_fire | | 107 | lens - camera - rupin - dang - goldberg | 37 | 107_lens_camera_rupin_dang | | 108 | xv - escaped - g3states - endif - define | 37 | 108_xv_escaped_g3states_endif | | 109 | mormons - jews - lds - sword - brigham | 36 | 109_mormons_jews_lds_sword | | 110 | resurrection - jesus - tomb - rise - luke | 36 | 110_resurrection_jesus_tomb_rise | | 111 | monitors - hours - nevai - day - monitor | 36 | 111_monitors_hours_nevai_day | | 112 | window - dialog - widget - xlib - application | 36 | 112_window_dialog_widget_xlib | | 113 | arrogance - truth - christians - arrogant - darren | 36 | 113_arrogance_truth_christians_arrogant | | 114 | gas - tear - unb - cs - jupiter | 36 | 114_gas_tear_unb_cs | | 115 | winfield - mattingly - peak - henderson - robinson | 35 | 115_winfield_mattingly_peak_henderson | | 116 | escrow - key - agencies - aclu - branch | 35 | 116_escrow_key_agencies_aclu | | 117 | judas - tyre - prophecy - prophecies - decenso | 35 | 117_judas_tyre_prophecy_prophecies | | 118 | image - processing - plplot - tools - analysis | 35 | 118_image_processing_plplot_tools | | 119 | eisa - isa - bus - vlb - motherboard | 35 | 119_eisa_isa_bus_vlb | | 120 | clipper - phone - phones - key - escrow | 35 | 120_clipper_phone_phones_key | | 121 | morris - team - jays - clemens - viola | 35 | 121_morris_team_jays_clemens | | 122 | space - moscow - shuttle - spaceflight - term | 34 | 122_space_moscow_shuttle_spaceflight | | 123 | hotel - voucher - ticket - hiram - airline | 34 | 123_hotel_voucher_ticket_hiram | | 124 | paint - wax - scratches - plastic - lisa | 34 | 124_paint_wax_scratches_plastic | | 125 | zeos - gateway - 486 - monitor - murthy | 34 | 125_zeos_gateway_486_monitor | | 126 | space - advertising - marketing - sky - billboard | 34 | 126_space_advertising_marketing_sky | | 127 | gopher - search - ftp - sites - exhibit | 34 | 127_gopher_search_ftp_sites | | 128 | 0d - _o - cx - c_ - 145 | 34 | 128_0d__o_cx_c_ | | 129 | gtoal - celp - speech - compression - toal | 33 | 129_gtoal_celp_speech_compression | | 130 | air - freon - aftermarket - behanna - r12 | 33 | 130_air_freon_aftermarket_behanna | | 131 | 3do - quicktime - ricardo - playback - mcmains | 33 | 131_3do_quicktime_ricardo_playback | | 132 | v4 - v6 - v8 - v12 - cdac | 33 | 132_v4_v6_v8_v12 | | 133 | font - fonts - character - truetype - windows | 33 | 133_font_fonts_character_truetype | | 134 | insurance - car - fault - rates - deductible | 32 | 134_insurance_car_fault_rates | | 135 | drivers - driver - card - jmarttila - actix | 32 | 135_drivers_driver_card_jmarttila | | 136 | tempest - holland - northeastern - utsa - cam | 32 | 136_tempest_holland_northeastern_utsa | | 137 | mustang - ford - camaro - howell - car | 32 | 137_mustang_ford_camaro_howell | | 138 | com4 - modem - com3 - port - 16550 | 31 | 138_com4_modem_com3_port | | 139 | deskjet - bubblejet - ink - printers - printer | 31 | 139_deskjet_bubblejet_ink_printers | | 140 | expose - window - event - buzz - main_win | 31 | 140_expose_window_event_buzz | | 141 | europeans - nhl - rauser - players - european | 31 | 141_europeans_nhl_rauser_players | | 142 | anonymous - privacy - anonymity - eff - internet | 31 | 142_anonymous_privacy_anonymity_eff | | 143 | vs - winner - bos - cal - chi | 31 | 143_vs_winner_bos_cal | | 144 | random - key - passwords - fifth - security | 31 | 144_random_key_passwords_fifth | | 145 | doctor - clinic - med - hoss - medicine | 31 | 145_doctor_clinic_med_hoss | | 146 | dc - shuttle - sdio - ssto - flight | 31 | 146_dc_shuttle_sdio_ssto | | 147 | split - newsgroup - cdrom - comp - graphics | 30 | 147_split_newsgroup_cdrom_comp | | 148 | nsa - cryptosystems - nea - paranoia - encryption | 30 | 148_nsa_cryptosystems_nea_paranoia | | 149 | colormap - visual - color - colormaps - dpy | 30 | 149_colormap_visual_color_colormaps | | 150 | jesus - brian - life - sandvik - kendig | 30 | 150_jesus_brian_life_sandvik | | 151 | atheism - asimov - timmons - alt - bake | 30 | 151_atheism_asimov_timmons_alt | | 152 | monitor - vga - monitors - lc - svga | 30 | 152_monitor_vga_monitors_lc | | 153 | eye - dominance - prk - handedness - rk | 29 | 153_eye_dominance_prk_handedness | | 154 | clinton - administration - qualcomm - tapped - drug | 29 | 154_clinton_administration_qualcomm_tapped | | 155 | fpu - c650 - coprocessor - 040 - 650 | 29 | 155_fpu_c650_coprocessor_040 | | 156 | cherry - coach - hockey - don - gilmour | 29 | 156_cherry_coach_hockey_don | | 157 | baptism - sin - aaron - baptized - infants | 29 | 157_baptism_sin_aaron_baptized | | 158 | car - dealer - price - sps - blue | 28 | 158_car_dealer_price_sps | | 159 | ir - dres - dnd - detector - detection | 28 | 159_ir_dres_dnd_detector | | 160 | rosicrucian - order - ch981 - amorc - tony | 28 | 160_rosicrucian_order_ch981_amorc | | 161 | health - tobacco - cesarean - cancer - smokeless | 28 | 161_health_tobacco_cesarean_cancer | | 162 | nt - windows - chicogo - os - rajiev | 28 | 162_nt_windows_chicogo_os | | 163 | king - kyle - adjective - nc - cramm | 28 | 163_king_kyle_adjective_nc | | 164 | muslims - serbs - croats - muslim - bosnian | 28 | 164_muslims_serbs_croats_muslim | | 165 | torre - hitter - gilkey - lankford - manager | 27 | 165_torre_hitter_gilkey_lankford | | 166 | bit - 24 - deniaud - bits - images | 27 | 166_bit_24_deniaud_bits | | 167 | dwi - infante - driving - drunk - speedy | 27 | 167_dwi_infante_driving_drunk | | 168 | xdm - server - login - graphic_display - error | 27 | 168_xdm_server_login_graphic_display | | 169 | 92 - hiv - aids - needles - 12 | 27 | 169_92_hiv_aids_needles | | 170 | diamond - stealth - drivers - card - speedstar | 27 | 170_diamond_stealth_drivers_card | | 171 | lopez - catchers - olson - braves - players | 27 | 171_lopez_catchers_olson_braves | | 172 | books - 02106 - 00 - chemistry - udel | 27 | 172_books_02106_00_chemistry | | 173 | duo - dock - apple - 230 - bredell | 27 | 173_duo_dock_apple_230 | | 174 | cable - antenna - tv - td - antennas | 27 | 174_cable_antenna_tv_td | | 175 | stadium - baseball - oswego - shea - mets | 26 | 175_stadium_baseball_oswego_shea | | 176 | images - image - geosphere - earth - unocal | 26 | 176_images_image_geosphere_earth | | 177 | sci - space - prado - henry - permanet | 26 | 177_sci_space_prado_henry | | 178 | peace - israel - palestinian - palestinians - talks | 26 | 178_peace_israel_palestinian_palestinians | | 179 | speed - x86 - 040 - 68040 - 680x0 | 26 | 179_speed_x86_040_68040 | | 180 | adcom - amp - amps - sound - microphone | 26 | 180_adcom_amp_amps_sound | | 181 | ati - ultra - drivers - gateway - 1280x1024 | 26 | 181_ati_ultra_drivers_gateway | | 182 | clipper - screw - chip - encryption - initiative | 26 | 182_clipper_screw_chip_encryption | | 183 | analog - seema - converter - hannover - 4066 | 26 | 183_analog_seema_converter_hannover | | 184 | mask - goalie - gtd597a - votes - hrivnak | 26 | 184_mask_goalie_gtd597a_votes | | 185 | 130 - rush - fast - lane - roads | 25 | 185_130_rush_fast_lane | | 186 | ashok - biochemistry - winqvt - kuleuven - liris | 25 | 186_ashok_biochemistry_winqvt_kuleuven | | 187 | room - summer - sublet - jhuvm - kitchen | 25 | 187_room_summer_sublet_jhuvm | | 188 | war - gulf - hussein - bombing - iraqi | 25 | 188_war_gulf_hussein_bombing | | 189 | ulf - erau - player - huot - shot | 25 | 189_ulf_erau_player_huot | | 190 | window - manager - xsizehints - bading - position | 25 | 190_window_manager_xsizehints_bading | | 191 | henrik - armenia - bm - planes - armenians | 25 | 191_henrik_armenia_bm_planes | | 192 | crypt - key - cryptography - des - ciphers | 25 | 192_crypt_key_cryptography_des | | 193 | amd - cyrix - 486dx2 - 486 - mhz | 25 | 193_amd_cyrix_486dx2_486 | | 194 | midi - sound - blaster - speaker - driver | 25 | 194_midi_sound_blaster_speaker | | 195 | mode - vga - tiang - svga - modes | 25 | 195_mode_vga_tiang_svga | | 196 | accelerations - acceleration - breathing - 45g - deaddio | 25 | 196_accelerations_acceleration_breathing_45g | | 197 | wire - wiring - ground - neutral - outlets | 24 | 197_wire_wiring_ground_neutral | | 198 | pain - bone - almanac - rib - massager | 24 | 198_pain_bone_almanac_rib | | 199 | reno - janet - madman - children - she | 24 | 199_reno_janet_madman_children | | 200 | barbecued - carcinogenic - meat - foods - risk | 24 | 200_barbecued_carcinogenic_meat_foods | | 201 | cmos - beeps - chimes - memory - error | 24 | 201_cmos_beeps_chimes_memory | | 202 | crohn - diet - ibd - inflammation - eat | 24 | 202_crohn_diet_ibd_inflammation | | 203 | wave - bikers - waved - cage - waving | 24 | 203_wave_bikers_waved_cage | | 204 | batf - warrant - knock - hallam - police | 24 | 204_batf_warrant_knock_hallam | | 205 | hacker - ethic - computer - hackers - programming | 23 | 205_hacker_ethic_computer_hackers | | 206 | mouse - motion - jumpy - smoothly - jump | 23 | 206_mouse_motion_jumpy_smoothly | | 207 | comet - jupiter - gehrels - sq - baalke | 23 | 207_comet_jupiter_gehrels_sq | | 208 | machines - precision - comments - contact - version | 23 | 208_machines_precision_comments_contact | | 209 | cosmo - angmar - alfalfa - pro - tsk | 23 | 209_cosmo_angmar_alfalfa_pro | | 210 | scsi - quadra - nodine - mac - cartridge | 23 | 210_scsi_quadra_nodine_mac | | 211 | adl - bullock - gerard - francisco - arens | 23 | 211_adl_bullock_gerard_francisco | | 212 | pgp - rsa - cryptography - code - patents | 23 | 212_pgp_rsa_cryptography_code | | 213 | koresh - sbc - backing - utarlg - enclosed | 23 | 213_koresh_sbc_backing_utarlg | | 214 | solvent - adhesive - duct - ruck - tape | 23 | 214_solvent_adhesive_duct_ruck | | 215 | command - spacecraft - galileo - baalke - timer | 23 | 215_command_spacecraft_galileo_baalke | | 216 | skin - dry - vaseline - rutin - acne | 23 | 216_skin_dry_vaseline_rutin | | 217 | gaza - gazans - ghetto - israeli - jews | 23 | 217_gaza_gazans_ghetto_israeli | | 218 | 03 - 02 - 04 - 01 - 05 | 22 | 218_03_02_04_01 | | 219 | ra - mormon - lds - bible - jesus | 22 | 219_ra_mormon_lds_bible | | 220 | abortion - child - fetus - margoli - abortions | 22 | 220_abortion_child_fetus_margoli | | 221 | 00 - wolverine - 1st - comics - hulk | 22 | 221_00_wolverine_1st_comics | | 222 | mac - 32 - os - stuffit - 800 | 22 | 222_mac_32_os_stuffit | | 223 | lyme - disease - fever - ld - infectious | 22 | 223_lyme_disease_fever_ld | | 224 | cobb - moral - morality - alexia - lis | 22 | 224_cobb_moral_morality_alexia | | 225 | sphere - den - p3 - p1 - p2 | 22 | 225_sphere_den_p3_p1 | | 226 | xputimage - shared - server - memory - animation | 22 | 226_xputimage_shared_server_memory | | 227 | rgb - luminosity - hue - red - green | 21 | 227_rgb_luminosity_hue_red | | 228 | pillion - riding - advice - passenger - ride | 21 | 228_pillion_riding_advice_passenger | | 229 | mouse - stuttgart - windows - driver - kasajian | 21 | 229_mouse_stuttgart_windows_driver | | 230 | gant - hirschbeck - umpire - strike - cox | 21 | 230_gant_hirschbeck_umpire_strike | | 231 | cursor - xterm - blinking - taylor - emu | 21 | 231_cursor_xterm_blinking_taylor | | 232 | tickets - 05pm - 35pm - june - ticket | 21 | 232_tickets_05pm_35pm_june | | 233 | ham - surges - alternator - interference - power | 21 | 233_ham_surges_alternator_interference | | 234 | marriage - married - ceremony - eyes - marry | 21 | 234_marriage_married_ceremony_eyes | | 235 | moa - bmw - rider - cactus - bmwmoa | 21 | 235_moa_bmw_rider_cactus | | 236 | number - phone - umass - ecs - line | 21 | 236_number_phone_umass_ecs | | 237 | bible - text - translations - texts - septuagint | 21 | 237_bible_text_translations_texts | | 238 | cop - officers - lmsc - lockheed - police | 21 | 238_cop_officers_lmsc_lockheed | | 239 | dxf - iff - format - autocad - pei | 20 | 239_dxf_iff_format_autocad | | 240 | roger - maynard - names - letter - laurentian | 20 | 240_roger_maynard_names_letter | | 241 | atheism - sapienza - atheists - fil - alt | 20 | 241_atheism_sapienza_atheists_fil | | 242 | video - verity - hdtv - compariators - input | 20 | 242_video_verity_hdtv_compariators | | 243 | yassin - deir - irgun - dir - village | 20 | 243_yassin_deir_irgun_dir | | 244 | god - predestination - saved - evil - grace | 20 | 244_god_predestination_saved_evil | | 245 | dialing - phones - tone - hugo - sweden | 20 | 245_dialing_phones_tone_hugo | | 246 | irq - interrupt - soundblaster - port - lpt1 | 20 | 246_irq_interrupt_soundblaster_port | | 247 | tongues - language - tounges - languages - koberg | 20 | 247_tongues_language_tounges_languages | | 248 | jsn104 - psuvm - hell - psu - damnation | 20 | 248_jsn104_psuvm_hell_psu | | 249 | chain - wax - behanna - maxima - cookson | 20 | 249_chain_wax_behanna_maxima | | 250 | bus - dx2 - 50mhz - dx - dx50 | 19 | 250_bus_dx2_50mhz_dx | | 251 | islamic - bcci - bank - jaeger - gregg | 19 | 251_islamic_bcci_bank_jaeger | | 252 | performa - lciii - iici - lc - pnet16 | 19 | 252_performa_lciii_iici_lc | | 253 | list - requests - bmw - request - mailing | 19 | 253_list_requests_bmw_request | | 254 | logo - rle - vgalogo - startup - lgo | 19 | 254_logo_rle_vgalogo_startup | | 255 | kidney - stones - calcium - she - stone | 19 | 255_kidney_stones_calcium_she | | 256 | phillies - phils - braves - wins - division | 19 | 256_phillies_phils_braves_wins | | 257 | monitor - lcd - screen - display - jiggles | 19 | 257_monitor_lcd_screen_display | | 258 | women - bobby - men - islamic - mozumder | 19 | 258_women_bobby_men_islamic | | 259 | ax - max - g9v - b8f - a86 | 19 | 259_ax_max_g9v_b8f | | 260 | koresh - mathew - bittrolff - david - risen | 19 | 260_koresh_mathew_bittrolff_david | | 261 | biggest - disappointment - smale - mvp - surprise | 19 | 261_biggest_disappointment_smale_mvp | | 262 | batf - oldham - blast - fokes - compound | 19 | 262_batf_oldham_blast_fokes | | 263 | sabbath - law - worship - paul - ceremonial | 19 | 263_sabbath_law_worship_paul | | 264 | joystick - joysticks - arcade - port - int15h | 19 | 264_joystick_joysticks_arcade_port | | 265 | captain - traded - captains - striped - resigned | 18 | 265_captain_traded_captains_striped | | 266 | mjm - fm - circuits - mixer - fsk | 18 | 266_mjm_fm_circuits_mixer | | 267 | cooling - towers - nuclear - plants - water | 18 | 267_cooling_towers_nuclear_plants | | 268 | she - were - her - apartment - they | 18 | 268_she_were_her_apartment | | 269 | pens - caps - eos - penguins - cdkaupan | 18 | 269_pens_caps_eos_penguins | | 270 | toyota - cruiser - suv - 4runner - cisco | 18 | 270_toyota_cruiser_suv_4runner | | 271 | love - god - dps - kodak - logic | 18 | 271_love_god_dps_kodak | | 272 | w4wg - network - workgroups - windows - lastdrive | 18 | 272_w4wg_network_workgroups_windows | | 273 | ticket - cop - speeding - chp - plates | 18 | 273_ticket_cop_speeding_chp | | 274 | lobby - sammons - ns111310 - colostate - letter | 18 | 274_lobby_sammons_ns111310_colostate | | 275 | ndw - spss - norton - ini - desktop | 18 | 275_ndw_spss_norton_ini | | 276 | uio - ifi - thomasp - parsli - quisling | 18 | 276_uio_ifi_thomasp_parsli | | 277 | motherboard - 386 - halcyon - 386dx - ruggiero | 18 | 277_motherboard_386_halcyon_386dx | | 278 | monitor - video - 610 - colors - screen | 18 | 278_monitor_video_610_colors | | 279 | oil - wd - 20w50 - 10w40 - militech | 18 | 279_oil_wd_20w50_10w40 | | 280 | printer - postscript - laser - laserjet - print | 18 | 280_printer_postscript_laser_laserjet | | 281 | probe - ford - car - newman - gt | 17 | 281_probe_ford_car_newman | | 282 | geico - insurance - davew - wonnacott - claim | 17 | 282_geico_insurance_davew_wonnacott | | 283 | 42 - tiff - philosophical - significance - joachim | 17 | 283_42_tiff_philosophical_significance | | 284 | omen - weight - fat - wa7kgx - forsberg | 17 | 284_omen_weight_fat_wa7kgx | | 285 | workspace - manager - managers - zip - workspaces | 17 | 285_workspace_manager_managers_zip | | 286 | fourd - vinge - vernor - 0565 - _the | 17 | 286_fourd_vinge_vernor_0565 | | 287 | mithras - pegasus - cunyvm - uoregon - magick | 17 | 287_mithras_pegasus_cunyvm_uoregon | | 288 | printer - adisak - pochanayon - pin - dot | 17 | 288_printer_adisak_pochanayon_pin | | 289 | gainey - bob - player - gilmour - maynard | 16 | 289_gainey_bob_player_gilmour | | 290 | adobe - photoshop - photo - platforms - shop | 16 | 290_adobe_photoshop_photo_platforms | | 291 | tank - tankbag - zipper - fj1100 - bgardner | 16 | 291_tank_tankbag_zipper_fj1100 | | 292 | disks - mac - 800k - binkley - 44mb | 16 | 292_disks_mac_800k_binkley | | 293 | graphics - pub - 128 - ray - rayshade | 16 | 293_graphics_pub_128_ray | | 294 | nubus - pds - lc - slot - marvin | 16 | 294_nubus_pds_lc_slot | | 295 | odometer - mileage - odometers - dealer - speedo | 16 | 295_odometer_mileage_odometers_dealer | | 296 | s1 - s2 - serial - key - unit | 16 | 296_s1_s2_serial_key | | 297 | lehigh - car - sports - ns1 - cars | 16 | 297_lehigh_car_sports_ns1 | | 298 | kjell - driver - hut - printer - backgrounder | 16 | 298_kjell_driver_hut_printer | | 299 | weapons - militia - weapon - foxvog - destruction | 16 | 299_weapons_militia_weapon_foxvog | | 300 | corn - seizures - paulson - seizure - cereals | 16 | 300_corn_seizures_paulson_seizure | | 301 | jagr - francis - minus - player - uvic | 16 | 301_jagr_francis_minus_player | | 302 | ingres - garrett - nixon - cambodia - tantrums | 16 | 302_ingres_garrett_nixon_cambodia | | 303 | 8051 - oscar - mont - 68hc16 - speth | 16 | 303_8051_oscar_mont_68hc16 | | 304 | tie - breaker - devils - islanders - record | 16 | 304_tie_breaker_devils_islanders | | 305 | motto - keith - caltech - pompous - schneider | 16 | 305_motto_keith_caltech_pompous | | 306 | ear - ears - ringing - earwax - vida | 16 | 306_ear_ears_ringing_earwax | | 307 | saturn - dealer - profit - sl2 - sc2 | 16 | 307_saturn_dealer_profit_sl2 | | 308 | tires - tire - fluids - abs - dot | 16 | 308_tires_tire_fluids_abs | | 309 | software - level - wingert - shuttle - process | 16 | 309_software_level_wingert_shuttle | | 310 | network - localtalk - ethernet - macs - appletalk | 16 | 310_network_localtalk_ethernet_macs | | 311 | mailing - list - bait - detweiler - rdetweil | 16 | 311_mailing_list_bait_detweiler | | 312 | satan - heaven - kicked - tyre - thou | 16 | 312_satan_heaven_kicked_tyre | | 313 | wip - sports - wfan - eagles - lupica | 15 | 313_wip_sports_wfan_eagles | | 314 | silence - moment - prayer - eeb1 - opposing | 15 | 314_silence_moment_prayer_eeb1 | | 315 | octopus - detroit - ice - hammerl - octopi | 15 | 315_octopus_detroit_ice_hammerl | | 316 | selective - borden - pork - service - abolish | 15 | 316_selective_borden_pork_service | | 317 | gajarsky - yogi - njin - stark - pilot | 15 | 317_gajarsky_yogi_njin_stark | | 318 | car - safety - centerline - saftey - collisions | 15 | 318_car_safety_centerline_saftey | | 319 | orion - film - prototype - henry - goltz | 15 | 319_orion_film_prototype_henry | | 320 | print - printer - file - claebaur - notepad | 15 | 320_print_printer_file_claebaur | | 321 | dod - denizens - kotl - doom - muck | 15 | 321_dod_denizens_kotl_doom | | 322 | display - remote - bielefeld - uphya001 - chooser | 15 | 322_display_remote_bielefeld_uphya001 | | 323 | spacecraft - funding - cuts - calpoly - digex | 15 | 323_spacecraft_funding_cuts_calpoly | | 324 | diesel - diesels - emissions - fuel - particulate | 15 | 324_diesel_diesels_emissions_fuel | | 325 | uva - partying - virginia - schools - beyer | 15 | 325_uva_partying_virginia_schools | | 326 | floptical - syquest - floppy - drives - floppies | 15 | 326_floptical_syquest_floppy_drives | | 327 | placebo - gr - roth - medicine - ron | 15 | 327_placebo_gr_roth_medicine | | 328 | canon - books - scripture - sirach - deuterocanonicals | 15 | 328_canon_books_scripture_sirach | | 329 | eliot - flat - boxer - 180 - v12 | 15 | 329_eliot_flat_boxer_180 | | 330 | firearms - smuggle - pound - guns - ban | 15 | 330_firearms_smuggle_pound_guns | | 331 | paradox - borland - quicken - sql - access | 15 | 331_paradox_borland_quicken_sql | | 332 | gun - buy - guns - stolen - buyback | 15 | 332_gun_buy_guns_stolen | | 333 | uranium - plutonium - nuclear - ryukoku - mccall | 15 | 333_uranium_plutonium_nuclear_ryukoku | | 334 | mosques - mosque - jerusalem - eggertj - jake | 15 | 334_mosques_mosque_jerusalem_eggertj | | 335 | clock - mhz - quadra - oscillator - centris | 15 | 335_clock_mhz_quadra_oscillator | | 336 | nixon - sternlight - mbeckman - crypto - strnlght | 15 | 336_nixon_sternlight_mbeckman_crypto | | 337 | african - workers - blacks - employees - crime | 15 | 337_african_workers_blacks_employees | | 338 | candida - vitamin - quack - pms - bloom | 14 | 338_candida_vitamin_quack_pms | | 339 | pluto - mission - alaska - probes - aurora | 14 | 339_pluto_mission_alaska_probes | | 340 | sabbath - salaris - black - lyrics - hell_2 | 14 | 340_sabbath_salaris_black_lyrics | | 341 | cd - rom - cdrom - adaptec - 3401 | 14 | 341_cd_rom_cdrom_adaptec | | 342 | fire - davidians - atf - fbi - napalm | 14 | 342_fire_davidians_atf_fbi | | 343 | drink - drinking - riding - ride - pnakada | 14 | 343_drink_drinking_riding_ride | | 344 | kubey - walks - obp - sac - hit | 14 | 344_kubey_walks_obp_sac | | 345 | cache - iisi - powercache - card - fpu | 14 | 345_cache_iisi_powercache_card | | 346 | murray - gm - quinn - vela - oakland | 14 | 346_murray_gm_quinn_vela | | 347 | simms - 256k - jh - cciw - csx | 14 | 347_simms_256k_jh_cciw | | 348 | 610 - centris - c610 - flaky - problems | 14 | 348_610_centris_c610_flaky | | 349 | cview - temp - moscom - zenkar - urc | 14 | 349_cview_temp_moscom_zenkar | | 350 | mhz - operational - clock - cpu - iisi | 14 | 350_mhz_operational_clock_cpu | | 351 | lock - locks - cobra - kryptonite - cable | 14 | 351_lock_locks_cobra_kryptonite | | 352 | wave - riceburner - squids - icomsim - squid | 14 | 352_wave_riceburner_squids_icomsim | | 353 | alarm - viper - alarms - sensor - car | 14 | 353_alarm_viper_alarms_sensor | | 354 | cubs - america - team - braves - talent | 14 | 354_cubs_america_team_braves | | 355 | pope - schism - church - catholic - sspx | 14 | 355_pope_schism_church_catholic | | 356 | christian - definition - christianity - jesus - christ | 14 | 356_christian_definition_christianity_jesus | | 357 | bonds - williams - batting - giants - punjabi | 14 | 357_bonds_williams_batting_giants | | 358 | bryce - arches - touring - dayton - fatcity | 14 | 358_bryce_arches_touring_dayton | | 359 | sound - stereo - channel - quadra - microphone | 14 | 359_sound_stereo_channel_quadra | | 360 | mormon - ceremonies - temple - temples - eusebius | 14 | 360_mormon_ceremonies_temple_temples | | 361 | reincarnation - elijah - karma - palo - gerry | 13 | 361_reincarnation_elijah_karma_palo | | 362 | fractal - fractals - compression - jr0930 - auckland | 13 | 362_fractal_fractals_compression_jr0930 | | 363 | marriage - marry - mormon - eternal - parents | 13 | 363_marriage_marry_mormon_eternal | | 364 | homeruns - boell - hit - hpcc01 - field | 13 | 364_homeruns_boell_hit_hpcc01 | | 365 | tv - flyback - exploding - prasad - emerson | 13 | 365_tv_flyback_exploding_prasad | | 366 | key - clarinet - tap - brad - proposal | 13 | 366_key_clarinet_tap_brad | | 367 | costly - memorial - museum - holocaust - techbook | 13 | 367_costly_memorial_museum_holocaust | | 368 | atm - fonts - tt - font - truetype | 13 | 368_atm_fonts_tt_font | | 369 | solder - boards - mask - green - silver | 13 | 369_solder_boards_mask_green | | 370 | temperature - henry - interstellar - sky - radiation | 13 | 370_temperature_henry_interstellar_sky | | 371 | answerfax - harris - rrrrr - select - wwerner | 13 | 371_answerfax_harris_rrrrr_select | | 372 | sale - suresh - mattress - table - rajaram | 13 | 372_sale_suresh_mattress_table | | 373 | handling - ntuvax - ntu - ba7116326 - handson | 13 | 373_handling_ntuvax_ntu_ba7116326 | | 374 | negev - bedouin - river - water - nysernet | 13 | 374_negev_bedouin_river_water | | 375 | cults - cult - muttiah - religions - religion | 13 | 375_cults_cult_muttiah_religions | | 376 | faith - saved - romans - lukewarm - deeds | 13 | 376_faith_saved_romans_lukewarm | | 377 | rh - liar - lunatic - he - bissell | 13 | 377_rh_liar_lunatic_he | | 378 | uart - 16550 - n5ial - uarts - modems | 13 | 378_uart_16550_n5ial_uarts | | 379 | rens - overreacting - dgbt - tapped - doc | 13 | 379_rens_overreacting_dgbt_tapped | | 380 | bible - language - commentary - christian - church | 13 | 380_bible_language_commentary_christian | | 381 | xclrp - mydisplay - palette_colors - drawindex - draw | 13 | 381_xclrp_mydisplay_palette_colors_drawindex | | 382 | oilers - pocklington - edmonton - northlands - yadallee | 13 | 382_oilers_pocklington_edmonton_northlands | | 383 | clinton - clipper - bush - rwing - pat | 13 | 383_clinton_clipper_bush_rwing | | 384 | easter - resurrection - celebration - pagan - goddess | 13 | 384_easter_resurrection_celebration_pagan | | 385 | ampere - amp - db - ohmite - company | 13 | 385_ampere_amp_db_ohmite | | 386 | logistician - 77 - wpi - ching - borque | 13 | 386_logistician_77_wpi_ching | | 387 | vram - simms - quadra - 512k - slots | 13 | 387_vram_simms_quadra_512k | | 388 | sin - hate - sinner - love - scott | 13 | 388_sin_hate_sinner_love | | 389 | prayers - jesus - prayer - jayne - husband | 12 | 389_prayers_jesus_prayer_jayne | | 390 | eridan - er1 - chuvashia - su - equip | 12 | 390_eridan_er1_chuvashia_su | | 391 | context - jim - joslin - meritt - mwunix | 12 | 391_context_jim_joslin_meritt | | 392 | mr2 - engine - eliot - noisy - shafts | 12 | 392_mr2_engine_eliot_noisy | | 393 | habitable - planets - atmosphere - oxygen - everest | 12 | 393_habitable_planets_atmosphere_oxygen | | 394 | sho - taurus - car - shifter - gk | 12 | 394_sho_taurus_car_shifter | | 395 | hall - fame - kingman - winfield - garvey | 12 | 395_hall_fame_kingman_winfield | | 396 | date - clock - dos - menu - stuck | 12 | 396_date_clock_dos_menu | | 397 | cd300i - umcc - apple - cdrom - cd | 12 | 397_cd300i_umcc_apple_cdrom | | 398 | beast - 666 - boylan - profile - usr | 12 | 398_beast_666_boylan_profile | | 399 | printer - imagewriter - appletalk - laserwriter - uchile | 12 | 399_printer_imagewriter_appletalk_laserwriter | | 400 | mpeg - quicktime - avi - melbourne - gregory | 12 | 400_mpeg_quicktime_avi_melbourne | | 401 | zarathushtra - magi - josephus - jesus - iranian | 12 | 401_zarathushtra_magi_josephus_jesus | | 402 | movies - bikes - csundh30 - cassidy - ursa | 12 | 402_movies_bikes_csundh30_cassidy | | 403 | satan - evil - lucifer - god - free | 12 | 403_satan_evil_lucifer_god | | 404 | solar - sail - sails - auburn - node | 12 | 404_solar_sail_sails_auburn | | 405 | limbaugh - rush - nlns - hitler - sahl | 12 | 405_limbaugh_rush_nlns_hitler | | 406 | warranty - techworks - credit - thacker - comtrade | 12 | 406_warranty_techworks_credit_thacker | | 407 | hiram - vhs - dk - kou - koutd | 12 | 407_hiram_vhs_dk_kou | | 408 | qur - koran - monash - bucaille - holy | 12 | 408_qur_koran_monash_bucaille | | 409 | bike - shipping - manish - ups - ship | 12 | 409_bike_shipping_manish_ups | | 410 | uniforms - marlins - lloyd - reds - mets | 12 | 410_uniforms_marlins_lloyd_reds | | 411 | rle - tga - povray - tmp - pov | 12 | 411_rle_tga_povray_tmp | | 412 | sunset - sunrise - drexel - cbis - rouben | 12 | 412_sunset_sunrise_drexel_cbis | | 413 | virtual - mfltd - sts - reality - vr | 11 | 413_virtual_mfltd_sts_reality | | 414 | ether - twist - mcaloon - dmcaloon - planets | 11 | 414_ether_twist_mcaloon_dmcaloon | | 415 | witnesses - trial - gm - new - judge | 11 | 415_witnesses_trial_gm_new | | 416 | disk - bios - drives - floppy - drive | 11 | 416_disk_bios_drives_floppy | | 417 | hook - phone - led - ring - hok | 11 | 417_hook_phone_led_ring | | 418 | pif - batch - bat - windows - environment | 11 | 418_pif_batch_bat_windows | | 419 | opel - manta - kadett - uiuc - gibbonsa | 11 | 419_opel_manta_kadett_uiuc | | 420 | winbench - winmarks - balog - diamond - stealth | 11 | 420_winbench_winmarks_balog_diamond | | 421 | iran - gulf - iranian - uae - iraq | 11 | 421_iran_gulf_iranian_uae | | 422 | voltage - current - supply - 12v - rooi | 11 | 422_voltage_current_supply_12v | | 423 | wrench - srb - thiokol - pliers - tool | 11 | 423_wrench_srb_thiokol_pliers | | 424 | xv - 24bit - image - 8bit - lilley | 11 | 424_xv_24bit_image_8bit | | 425 | baptists - trincoll - banging - sociopaths - marrying | 11 | 425_baptists_trincoll_banging_sociopaths | | 426 | jb - diabetes - ron - roth - anello | 11 | 426_jb_diabetes_ron_roth | | 427 | jesus - commandments - god - law - commandment | 11 | 427_jesus_commandments_god_law | | 428 | hitler - nazis - roehm - chancellor - nazi | 11 | 428_hitler_nazis_roehm_chancellor | | 429 | freemasonry - masonry - masonic - baptist - southern | 11 | 429_freemasonry_masonry_masonic_baptist | | 430 | cd300 - bauer - cd - multisession - toshiba | 11 | 430_cd300_bauer_cd_multisession | | 431 | x11r5 - xsun - o_rdonly - fonts - 0666 | 11 | 431_x11r5_xsun_o_rdonly_fonts | | 432 | controller - ide - bus - fdd - sec | 11 | 432_controller_ide_bus_fdd | | 433 | gusto - heart - cardiac - uts - pvc | 11 | 433_gusto_heart_cardiac_uts | | 434 | licensed - 2a42dubinski - carlos - change - hex | 11 | 434_licensed_2a42dubinski_carlos_change | | 435 | convertible - wife - targa - wants - car | 11 | 435_convertible_wife_targa_wants | | 436 | scores - posts - savoy - brock - hernandez | 11 | 436_scores_posts_savoy_brock | | 437 | lcd - malouf - monitor - damico - projector | 10 | 437_lcd_malouf_monitor_damico | | 438 | dtr - rts - dsr - cts - dce | 10 | 438_dtr_rts_dsr_cts | | 439 | 2600 - atari - tia - 5200 - 4k | 10 | 439_2600_atari_tia_5200 | | 440 | rs232 - ttl - ka3uww - loopback - ic | 10 | 440_rs232_ttl_ka3uww_loopback | | 441 | contradictions - medtronic - archer - skiba - biblical | 10 | 441_contradictions_medtronic_archer_skiba | | 442 | princeton - fester - black - roger - lazy | 10 | 442_princeton_fester_black_roger | | 443 | wordbasic - filenames - format - file - word | 10 | 443_wordbasic_filenames_format_file | </details> ## Training hyperparameters * calculate_probabilities: False * language: None * low_memory: False * min_topic_size: 10 * n_gram_range: (1, 1) * nr_topics: None * seed_topic_list: None * top_n_words: 10 * verbose: False * zeroshot_min_similarity: 0.7 * zeroshot_topic_list: None ## Framework versions * Numpy: 1.25.2 * HDBSCAN: 0.8.33 * UMAP: 0.5.6 * Pandas: 2.0.3 * Scikit-Learn: 1.2.2 * Sentence-transformers: 2.7.0 * Transformers: 4.40.1 * Numba: 0.58.1 * Plotly: 5.15.0 * Python: 3.10.12
SinniDcat/LLama3_chinese_instrument_unstable
SinniDcat
2024-05-08T08:45:39Z
1
0
null
[ "gguf", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
null
2024-05-08T07:59:16Z
--- license: apache-2.0 ---
isinTarik/model_16_32
isinTarik
2024-05-08T08:45:08Z
105
0
transformers
[ "transformers", "pytorch", "safetensors", "gemma", "text-generation", "unsloth", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-07T09:02:20Z
--- tags: - unsloth --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
ikeno-ada/gemma-1.1-2b-it-Quanto-4bit
ikeno-ada
2024-05-08T08:43:52Z
106
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "region:us" ]
text-generation
2024-05-08T08:11:59Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
youngsangroh/whisper-small-finetuned-atco2-asr-atcosim
youngsangroh
2024-05-08T08:43:06Z
89
1
transformers
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "generated_from_trainer", "en", "dataset:jlvdoorn/atco2-asr-atcosim", "base_model:openai/whisper-small", "base_model:finetune:openai/whisper-small", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2024-05-08T05:52:29Z
--- language: - en license: apache-2.0 base_model: openai/whisper-small tags: - generated_from_trainer datasets: - jlvdoorn/atco2-asr-atcosim metrics: - wer model-index: - name: Whisper Small En - Whisper with atco2-asr-atcosim results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: 'This is a dataset constructed from two datasets: ATCO2-ASR and ATCOSIM.' type: jlvdoorn/atco2-asr-atcosim args: 'config: en, split: test' metrics: - name: Wer type: wer value: 0.02577651759247326 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Whisper Small En - Whisper with atco2-asr-atcosim This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the This is a dataset constructed from two datasets: ATCO2-ASR and ATCOSIM. dataset. It achieves the following results on the evaluation set: - Loss: 0.0010 - Wer: 0.0258 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 4000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:------:|:----:|:---------------:|:------:| | 0.0637 | 1.9763 | 1000 | 0.0962 | 7.4365 | | 0.0154 | 3.9526 | 2000 | 0.0163 | 2.3972 | | 0.002 | 5.9289 | 3000 | 0.0027 | 1.5015 | | 0.0003 | 7.9051 | 4000 | 0.0010 | 0.0258 | ### Framework versions - Transformers 4.40.2 - Pytorch 2.2.1+cu121 - Datasets 2.19.1 - Tokenizers 0.19.1
PetroGPT/Breeze-Petro-7B-Instruct-v1
PetroGPT
2024-05-08T08:42:06Z
7
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "chemistry", "code", "text-generation-inference", "conversational", "en", "zh", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-04-23T09:20:44Z
--- library_name: transformers tags: - chemistry - code - text-generation-inference license: apache-2.0 language: - en - zh metrics: - accuracy - code_eval --- # Breeze-Petro-7B-Instruct-v1 - Model creator: [MediaTek Research](https://huggingface.co/MediaTek-Research) - Original model: [MediaTek-Research/Breeze-7B-Instruct-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-Instruct-v1_0) ### Model Description This is a model fine-tuned based on Breeze-7B-Instruct-v1_0. The training set is mainly based on chemical knowledge and procedural knowledge. Supplement knowledge about the petroleum industry. - **Developed by:** RebeccaChou - **License:** apache-2.0 - **Finetuned from model :** [MediaTek-Research/Breeze-7B-Instruct-v1_0 - **Language(s) (NLP):** [English.繁體中文] # 📖 Table of Contents 1.[Open LLM Leaderboard](#🏆-open-llm-leaderboard) - ARC - HellaSwag - MMLU - TruthfulQA - Winogrande - GSM8K 3. [EvalPlus Leaderboard](#⚡-evalplus-leaderboard) - HumanEval - HumanEval_Plus - MBPP - MBPP_Plus 4. [Prompt Format](#⚗️-prompt-format) 5. [Quantized Models](#🛠️-quantized-models) 6. [Gratitude](#🙏-gratitude) ## 🏆 Open LLM Leaderboard WestSeverus-7B-DPO-v2 is one of the top 7B model in Open LLM Leaderboard and it outperforms on TruthfulQA and GSM8K. | Metric |Value| |---------------------------------|----:| |Avg. |59.32| |AI2 Reasoning Challenge (25-Shot)|58.87| |HellaSwag (10-Shot) |79.17| |MMLU (5-Shot) |56.62| |TruthfulQA (0-shot) |46.36| |Winogrande (5-shot) |73.64| |GSM8k (5-shot) |41.24| Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Rebecca19990101__Breeze-Petro-7B-Instruct-v1) ## ⚡ EvalPlus Leaderboard | Model | HumanEval | HumanEval_Plus| MBPP | MBPP_Plus | |---|---:|---:|---:|---:| | phi-2-2.7B |48.2|43.3|61.9|51.4| | | | | | | | SOLAR-10.7B-Instruct-v1.0 | 42.1 | 34.3 | 42.9 | 34.6 | | CodeLlama-7B| 37.8| 34.1 | 57.6 |45.4 | ## 🛠️ Quantized Models * **GGUF** https://huggingface.co/Rebecca19990101/breeze-petro-7b-instruct-v1-q4_k_m.gguf/tree/main ## Training Details ### Training Data - **Dataset:** Rebecca19990101/petro-dataset-v2
Elizezen/Zero-dialogue-for-Mistral-Based-Japanese-LLM
Elizezen
2024-05-08T08:41:50Z
0
0
null
[ "safetensors", "license:apache-2.0", "region:us" ]
null
2024-05-07T04:51:21Z
--- license: apache-2.0 --- text-generation-webuiなどで使えるLoRA (ほぼ)地の文のみを出力させます。AIのべりすとの「ナラティブ」スタイルみたいなイメージ
Mag0g/Ezekiel25_17
Mag0g
2024-05-08T08:39:43Z
131
0
transformers
[ "transformers", "safetensors", "stablelm", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T08:32:52Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
jsingh/autoflow-math-v0.3
jsingh
2024-05-08T08:37:30Z
0
1
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-04-27T00:32:14Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
MMEX/QLora_llama2-hc
MMEX
2024-05-08T08:36:34Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-05-08T08:36:27Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
4season/sft_model_test1
4season
2024-05-08T08:30:00Z
9
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "en", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T07:16:03Z
--- license: apache-2.0 language: - en --- # 4season/sft_model_test1 # **Introduction** This model is test version, sft model. We utilize state-of-the-art instruction fine-tuning methods. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - gradient_accumulation_steps: 8 - total_train_batch_size: 128 - total_eval_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - num_epochs: 1.0
huynq3Cyradar/bert-large-finetuned-phishing-webpage-version
huynq3Cyradar
2024-05-08T08:26:01Z
110
0
transformers
[ "transformers", "tensorboard", "safetensors", "bert", "text-classification", "generated_from_trainer", "base_model:google-bert/bert-base-multilingual-cased", "base_model:finetune:google-bert/bert-base-multilingual-cased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-06T09:43:34Z
--- license: apache-2.0 base_model: google-bert/bert-base-multilingual-cased tags: - generated_from_trainer metrics: - accuracy - precision - recall model-index: - name: bert-large-finetuned-phishing-webpage-version results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-large-finetuned-phishing-webpage-version This model is a fine-tuned version of [google-bert/bert-base-multilingual-cased](https://huggingface.co/google-bert/bert-base-multilingual-cased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.2062 - Accuracy: 0.9188 - Precision: 0.9517 - Recall: 0.8689 - False Positive Rate: 0.0381 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 256 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | False Positive Rate | |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:-------------------:| | No log | 1.0 | 394 | 0.2675 | 0.8918 | 0.9680 | 0.7926 | 0.0226 | | 0.3256 | 2.0 | 788 | 0.2225 | 0.9124 | 0.9640 | 0.8424 | 0.0272 | | 0.2008 | 3.0 | 1182 | 0.2062 | 0.9188 | 0.9517 | 0.8689 | 0.0381 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.1.2 - Datasets 2.18.0 - Tokenizers 0.15.2
imagepipeline/FACE
imagepipeline
2024-05-08T08:21:41Z
0
0
null
[ "imagepipeline", "imagepipeline.io", "text-to-image", "ultra-realistic", "license:creativeml-openrail-m", "region:us" ]
text-to-image
2024-05-08T08:21:39Z
--- license: creativeml-openrail-m tags: - imagepipeline - imagepipeline.io - text-to-image - ultra-realistic pinned: false pipeline_tag: text-to-image --- ## FACE <img src="https://via.placeholder.com/468x300?text=App+Screenshot+Here" alt="Generated on Image Pipeline" style="border-radius: 10px;"> **This lora model is uploaded on [imagepipeline.io](https://imagepipeline.io/)** Model details - FACESITTING [![Try this model](https://img.shields.io/badge/try_this_model-image_pipeline-BD9319)](https://imagepipeline.io/models/FACE?id=00b62a0f-ad42-4798-be36-aa77bbb6e035/) ## How to try this model ? You can try using it locally or send an API call to test the output quality. Get your `API_KEY` from [imagepipeline.io](https://imagepipeline.io/). No payment required. Coding in `php` `javascript` `node` etc ? Checkout our documentation [![documentation](https://img.shields.io/badge/documentation-image_pipeline-blue)](https://docs.imagepipeline.io/docs/introduction) ```python import requests import json url = "https://imagepipeline.io/sdxl/text2image/v1/run" payload = json.dumps({ "model_id": "sdxl", "prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K", "negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime", "width": "512", "height": "512", "samples": "1", "num_inference_steps": "30", "safety_checker": false, "guidance_scale": 7.5, "multi_lingual": "no", "embeddings": "", "lora_models": "00b62a0f-ad42-4798-be36-aa77bbb6e035", "lora_weights": "0.5" }) headers = { 'Content-Type': 'application/json', 'API-Key': 'your_api_key' } response = requests.request("POST", url, headers=headers, data=payload) print(response.text) } ``` Get more ready to use `MODELS` like this for `SD 1.5` and `SDXL` : [![All models](https://img.shields.io/badge/Get%20All%20Models-image_pipeline-BD9319)](https://imagepipeline.io/models) ### API Reference #### Generate Image ```http https://api.imagepipeline.io/sdxl/text2image/v1 ``` | Headers | Type | Description | |:----------------------| :------- |:-------------------------------------------------------------------------------------------------------------------| | `API-Key` | `str` | Get your `API_KEY` from [imagepipeline.io](https://imagepipeline.io/) | | `Content-Type` | `str` | application/json - content type of the request body | | Parameter | Type | Description | | :-------- | :------- | :------------------------- | | `model_id` | `str` | Your base model, find available lists in [models page](https://imagepipeline.io/models) or upload your own| | `prompt` | `str` | Text Prompt. Check our [Prompt Guide](https://docs.imagepipeline.io/docs/SD-1.5/docs/extras/prompt-guide) for tips | | `num_inference_steps` | `int [1-50]` | Noise is removed with each step, resulting in a higher-quality image over time. Ideal value 30-50 (without LCM) | | `guidance_scale` | `float [1-20]` | Higher guidance scale prioritizes text prompt relevance but sacrifices image quality. Ideal value 7.5-12.5 | | `lora_models` | `str, array` | Pass the model_id(s) of LoRA models that can be found in models page | | `lora_weights` | `str, array` | Strength of the LoRA effect | --- license: creativeml-openrail-m tags: - imagepipeline - imagepipeline.io - text-to-image - ultra-realistic pinned: false pipeline_tag: text-to-image --- ### Feedback If you have any feedback, please reach out to us at hello@imagepipeline.io #### 🔗 Visit Website [![portfolio](https://img.shields.io/badge/image_pipeline-BD9319?style=for-the-badge&logo=gocd&logoColor=white)](https://imagepipeline.io/) If you are the original author of this model, please [click here](https://airtable.com/apprTaRnJbDJ8ufOx/shr4g7o9B6fWfOlUR) to add credits
ameerazam08/MuseTalk
ameerazam08
2024-05-08T08:20:48Z
0
3
diffusers
[ "diffusers", "onnx", "safetensors", "license:creativeml-openrail-m", "region:us" ]
null
2024-05-08T08:03:08Z
--- title: MuseTalkDemo emoji: 🌍 colorFrom: gray colorTo: purple sdk: docker pinned: false license: creativeml-openrail-m app_file: app.py app_port: 7860 --- ALL Setup for MuseTalk Clone and Run ``` Build environment We recommend a python version >=3.10 and cuda version =11.7. Then build environment as follows: pip install -r requirements.txt mmlab packages pip install --no-cache-dir -U openmim mim install mmengine mim install "mmcv>=2.0.1" mim install "mmdet>=3.1.0" mim install "mmpose>=1.1.0" Download ffmpeg-static Download the ffmpeg-static and export FFMPEG_PATH=/path/to/ffmpeg for example: export FFMPEG_PATH=/musetalk/ffmpeg-4.4-amd64-static Download weights You can download weights manually as follows: Download our trained weights. Download the weights of other components: sd-vae-ft-mse whisper dwpose face-parse-bisent resnet18 Finally, these weights should be organized in models as follows: ./models/ ├── musetalk │ └── musetalk.json │ └── pytorch_model.bin ├── dwpose │ └── dw-ll_ucoco_384.pth ├── face-parse-bisent │ ├── 79999_iter.pth │ └── resnet18-5c106cde.pth ├── sd-vae-ft-mse │ ├── config.json │ └── diffusion_pytorch_model.bin └── whisper └── tiny.pt Quickstart Inference Here, we provide the inference script. python -m scripts.inference --inference_config configs/inference/test.yaml configs/inference/test.yaml is the path to the inference configuration file, including video_path and audio_path. The video_path should be either a video file, an image file or a directory of images. You are recommended to input video with 25fps, the same fps used when training the model. If your video is far less than 25fps, you are recommended to apply frame interpolation or directly convert the video to 25fps using ffmpeg. Use of bbox_shift to have adjustable results 🔎 We have found that upper-bound of the mask has an important impact on mouth openness. Thus, to control the mask region, we suggest using the bbox_shift parameter. Positive values (moving towards the lower half) increase mouth openness, while negative values (moving towards the upper half) decrease mouth openness. You can start by running with the default configuration to obtain the adjustable value range, and then re-run the script within this range. For example, in the case of Xinying Sun, after running the default configuration, it shows that the adjustable value rage is [-9, 9]. Then, to decrease the mouth openness, we set the value to be -7. python -m scripts.inference --inference_config configs/inference/test.yaml --bbox_shift -7 📌 More technical details can be found in bbox_shift. Combining MuseV and MuseTalk As a complete solution to virtual human generation, you are suggested to first apply MuseV to generate a video (text-to-video, image-to-video or pose-to-video) by referring this. Frame interpolation is suggested to increase frame rate. Then, you can use MuseTalk to generate a lip-sync video by referring this. 🆕 Real-time inference Here, we provide the inference script. This script first applies necessary pre-processing such as face detection, face parsing and VAE encode in advance. During inference, only UNet and the VAE decoder are involved, which makes MuseTalk real-time. python -m scripts.realtime_inference --inference_config configs/inference/realtime.yaml --batch_size 4 configs/inference/realtime.yaml is the path to the real-time inference configuration file, including preparation, video_path , bbox_shift and audio_clips. Set preparation to True in realtime.yaml to prepare the materials for a new avatar. (If the bbox_shift has changed, you also need to re-prepare the materials.) After that, the avatar will use an audio clip selected from audio_clips to generate video. Inferring using: data/audio/yongen.wav While MuseTalk is inferring, sub-threads can simultaneously stream the results to the users. The generation process can achieve 30fps+ on an NVIDIA Tesla V100. Set preparation to False and run this script if you want to genrate more videos using the same avatar. Note for Real-time inference If you want to generate multiple videos using the same avatar/video, you can also use this script to SIGNIFICANTLY expedite the generation process. In the previous script, the generation time is also limited by I/O (e.g. saving images). If you just want to test the generation speed without saving the images, you can run python -m scripts.realtime_inference --inference_config configs/inference/realtime.yaml --skip_save_images ```
dinhhung1508/Seallm-7b-v2.5-summary-vietnamese-article-v1-gguf
dinhhung1508
2024-05-08T08:20:35Z
13
0
transformers
[ "transformers", "gguf", "gemma", "text-generation-inference", "unsloth", "en", "base_model:SeaLLMs/SeaLLM-7B-v2.5", "base_model:quantized:SeaLLMs/SeaLLM-7B-v2.5", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
null
2024-05-08T08:18:45Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - gemma - gguf base_model: SeaLLMs/SeaLLM-7B-v2.5 --- # Uploaded model - **Developed by:** dinhhung1508 - **License:** apache-2.0 - **Finetuned from model :** SeaLLMs/SeaLLM-7B-v2.5 This gemma model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
UnicomLLM/Unichat-llama3-Chinese-8B-gguf
UnicomLLM
2024-05-08T08:19:14Z
9
2
null
[ "gguf", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
null
2024-04-30T06:16:49Z
--- license: apache-2.0 --- This is the quantization version of Unichat-llama3-Chinese-8B model.
Sumail/Chalice15
Sumail
2024-05-08T08:17:07Z
131
0
transformers
[ "transformers", "safetensors", "stablelm", "text-generation", "mergekit", "merge", "conversational", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T08:15:35Z
--- base_model: - vapegod/stable5 - vapegod/stable library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the SLERP merge method. ### Models Merged The following models were included in the merge: * [vapegod/stable5](https://huggingface.co/vapegod/stable5) * [vapegod/stable](https://huggingface.co/vapegod/stable) ### Configuration The following YAML configuration was used to produce this model: ```yaml slices: - sources: - model: vapegod/stable5 layer_range: [0, 24] - model: vapegod/stable layer_range: [0, 24] merge_method: slerp base_model: vapegod/stable5 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ```
opencsg/csg-wukong-1B-sft-bf16
opencsg
2024-05-08T08:15:23Z
151
1
transformers
[ "transformers", "safetensors", "llama", "text-generation", "code", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-04-30T14:48:46Z
--- language: - en pipeline_tag: text-generation tags: - code license: apache-2.0 --- # **csg-wukong-1B-sft-bf16** [[中文]](#chinese) [[English]](#english) <a id="english"></a> <p align="center"> <img width="900px" alt="OpenCSG" src="./csg-wukong-logo-green.jpg"> </p> <p align="center"><a href="https://portal.opencsg.com/models">[OpenCSG Community]</a> <a href="hhttps://github.com/OpenCSGs/Awesome-SLMs">[github]</a> <a href="https://cdn-uploads.huggingface.co/production/uploads/64c71b27d43e4dee51a8b31a/HU6vz21qKTEmUBCWqCFh9.jpeg">[wechat]</a> <a href="https://twitter.com/OpenCsg">[Twitter]</a> </p> </div> OpenCSG stands for Converged resources, Software refinement, and Generative LM. The 'C' represents Converged resources, indicating the integration and full utilization of hybrid resources. The 'S' stands for Software refinement, signifying software that is refined by large models. The 'G' represents Generative LM, which denotes widespread, inclusive, and democratized generative large models. The vision of OpenCSG is to empower every industry, every company, and every individual to own their models. We adhere to the principles of openness and open source, making the large model software stack of OpenCSG available to the community. We welcome everyone to use, send feedback, and contribute collaboratively. ## Model Description **csg-wukong-1B-sft-bf16** was finetuned on [csg-wukong-1B](https://huggingface.co/opencsg/csg-wukong-1B). <br> we will introduce more information about csg-wukong-1B. ## Model Evaluation results We submitted csg-wukong-1B on the [open_llm_leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard), and the results show our model ranked the 8th among the ~1.5B pretrained small language models. ![image/png](https://cdn-uploads.huggingface.co/production/uploads/661790397437201d78141856/_HRTxL6N0qnNPNt-P8k9k.png) # Training ## Hardware - **GPUs:** 16 H800 - **Training time:** 43days ## Software - **Orchestration:** [Deepspeed](https://github.com/OpenCSGs) - **Neural networks:** [PyTorch](https://github.com/pytorch/pytorch) - **BP16 if applicable:** [apex](https://github.com/NVIDIA/apex) <a id="chinese"></a> <p> </p> # OpenCSG介绍 <p align="center"> <img width="300px" alt="OpenCSG" src="https://cdn-uploads.huggingface.co/production/uploads/64c71b27d43e4dee51a8b31a/GwYXPKuEoGCGcMICeW-sb.jpeg"> </p> <p align="center"><a href="https://opencsg.com/models">[OpenCSG 社区]</a> <a href="https://github.com/OpenCSGs/Awesome-SLMs">[github]</a> <a href="https://cdn-uploads.huggingface.co/production/uploads/64c71b27d43e4dee51a8b31a/HU6vz21qKTEmUBCWqCFh9.jpeg">[微信]</a> <a href="https://twitter.com/OpenCsg">[推特]</a> </p> </div> OpenCSG中 Open是开源开放;C 代表 Converged resources,整合和充分利用的混合异构资源优势,算力降本增效;S 代表 Software refined,重新定义软件的交付方式,通过大模型驱动软件开发,人力降本增效;G 代表 Generative LM,大众化、普惠化和民主化的可商用的开源生成式大模型。 OpenCSG的愿景是让每个行业、每个公司、每个人都拥有自己的模型。 我们坚持开源开放的原则,将OpenCSG的大模型软件栈开源到社区,欢迎使用、反馈和参与共建,欢迎关注。 ## 模型介绍 **csg-wukong-1B-sft-bf16** 在[csg-wukong-1B](https://huggingface.co/opencsg/csg-wukong-1B)预训练模型上微调而成. <br> 我们将在后面介绍更多关于这个模型的信息。 ## 模型评测结果 我们把csg-wukong-1B模型提交到[open_llm_leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)榜单上,结果显示我们的模型目前在~1.5B小语言模型中排名第8。 ![image/png](https://cdn-uploads.huggingface.co/production/uploads/661790397437201d78141856/ZfWZ1Fd7ccKrJVx0okV9z.png) # 训练 ## 硬件资源 - **GPU数量:** 16 H800 - **训练时间:** 43天 ## 软件使用 - **微调训练框架:** [Deepspeed](https://github.com/OpenCSGs) - **深度学习框架:** [PyTorch](https://github.com/pytorch/pytorch) - **BP16:** [apex](https://github.com/NVIDIA/apex)
MohitML/gpt2-reuters_tokenizer
MohitML
2024-05-08T08:14:12Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-05-08T08:14:11Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
bllossom-advanced/bllossom-llama-3-8b-65k-base
bllossom-advanced
2024-05-08T08:09:44Z
7
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T07:46:33Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
iguanaYu/distilroberta-base-finetuned-wikitext2
iguanaYu
2024-05-08T08:08:26Z
163
0
transformers
[ "transformers", "tensorboard", "safetensors", "roberta", "fill-mask", "generated_from_trainer", "base_model:distilbert/distilroberta-base", "base_model:finetune:distilbert/distilroberta-base", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2024-05-08T07:41:22Z
--- license: apache-2.0 base_model: distilroberta-base tags: - generated_from_trainer model-index: - name: distilroberta-base-finetuned-wikitext2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilroberta-base-finetuned-wikitext2 This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.8611 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.0841 | 1.0 | 2406 | 1.9362 | | 1.9866 | 2.0 | 4812 | 1.8845 | | 1.9442 | 3.0 | 7218 | 1.8355 | ### Framework versions - Transformers 4.40.1 - Pytorch 2.2.1+cu121 - Datasets 2.19.1 - Tokenizers 0.19.1
gaianet/Nomic-embed-text-v1.5-Embedding-GGUF
gaianet
2024-05-08T08:04:43Z
35,640
5
null
[ "gguf", "license:apache-2.0", "endpoints_compatible", "region:us", "feature-extraction" ]
null
2024-05-08T07:49:18Z
--- license: apache-2.0 ---
aaron-di/YamshadowExperiment28-7B-Ties
aaron-di
2024-05-08T08:04:12Z
6
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "automerger", "base_model:automerger/YamShadow-7B", "base_model:merge:automerger/YamShadow-7B", "base_model:yam-peleg/Experiment28-7B", "base_model:merge:yam-peleg/Experiment28-7B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T08:00:49Z
--- license: apache-2.0 tags: - merge - mergekit - lazymergekit - automerger base_model: - automerger/YamShadow-7B - yam-peleg/Experiment28-7B --- ## 🧩 Configuration ```yaml models: - model: automerger/YamShadow-7B parameters: density: 0.5 weight: 0.5 - model: yam-peleg/Experiment28-7B parameters: density: 0.5 weight: 0.5 merge_method: ties base_model: automerger/YamShadow-7B dtype: float16 random_seed: 0 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "aaron-di/YamshadowExperiment28-7B-Ties" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
NewsLLM/llama-3-8b-NewsLLM-phase1
NewsLLM
2024-05-08T08:02:38Z
9
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T07:54:40Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
dinhhung1508/Seallm-7b-v2.5-summary-vietnamese-article-v1
dinhhung1508
2024-05-08T07:59:36Z
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "gemma", "trl", "en", "base_model:SeaLLMs/SeaLLM-7B-v2.5", "base_model:finetune:SeaLLMs/SeaLLM-7B-v2.5", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2024-05-08T07:59:34Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - gemma - trl base_model: SeaLLMs/SeaLLM-7B-v2.5 --- # Uploaded model - **Developed by:** dinhhung1508 - **License:** apache-2.0 - **Finetuned from model :** SeaLLMs/SeaLLM-7B-v2.5 This gemma model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
dinhhung1508/Seallm-7b-v2.5-summary-vietnamese-article-v1-merged_4bit
dinhhung1508
2024-05-08T07:57:06Z
81
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "base_model:SeaLLMs/SeaLLM-7B-v2.5", "base_model:quantized:SeaLLMs/SeaLLM-7B-v2.5", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
2024-05-08T07:55:32Z
--- language: - en license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - gemma - trl base_model: SeaLLMs/SeaLLM-7B-v2.5 --- # Uploaded model - **Developed by:** dinhhung1508 - **License:** apache-2.0 - **Finetuned from model :** SeaLLMs/SeaLLM-7B-v2.5 This gemma model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
SinniDcat/LLAMA3-chnese-instrument-test-lora_model
SinniDcat
2024-05-08T07:54:15Z
0
0
null
[ "safetensors", "license:apache-2.0", "region:us" ]
null
2024-05-08T07:47:54Z
--- license: apache-2.0 ---
yweslakarep/vit-base-patch16-224-in21k-finetuned-lora-food101
yweslakarep
2024-05-08T07:52:37Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-05-08T07:52:35Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
chibeenot/lora_model_test
chibeenot
2024-05-08T07:52:30Z
1
0
peft
[ "peft", "arxiv:1910.09700", "base_model:ai-forever/FRED-T5-1.7B", "base_model:adapter:ai-forever/FRED-T5-1.7B", "region:us" ]
null
2024-05-08T06:26:45Z
--- library_name: peft base_model: ai-forever/FRED-T5-1.7B --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.7.1
ahmed-kh/superhero
ahmed-kh
2024-05-08T07:50:12Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2024-05-08T07:50:12Z
--- license: apache-2.0 ---
hustvl/yolos-small
hustvl
2024-05-08T07:49:12Z
49,030
61
transformers
[ "transformers", "pytorch", "safetensors", "yolos", "object-detection", "vision", "dataset:coco", "arxiv:2106.00666", "license:apache-2.0", "endpoints_compatible", "region:us" ]
object-detection
2022-04-26T09:38:22Z
--- license: apache-2.0 tags: - object-detection - vision datasets: - coco widget: - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/savanna.jpg example_title: Savanna - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/football-match.jpg example_title: Football Match - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/airport.jpg example_title: Airport --- # YOLOS (small-sized) model YOLOS model fine-tuned on COCO 2017 object detection (118k annotated images). It was introduced in the paper [You Only Look at One Sequence: Rethinking Transformer in Vision through Object Detection](https://arxiv.org/abs/2106.00666) by Fang et al. and first released in [this repository](https://github.com/hustvl/YOLOS). Disclaimer: The team releasing YOLOS did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description YOLOS is a Vision Transformer (ViT) trained using the DETR loss. Despite its simplicity, a base-sized YOLOS model is able to achieve 42 AP on COCO validation 2017 (similar to DETR and more complex frameworks such as Faster R-CNN). The model is trained using a "bipartite matching loss": one compares the predicted classes + bounding boxes of each of the N = 100 object queries to the ground truth annotations, padded up to the same length N (so if an image only contains 4 objects, 96 annotations will just have a "no object" as class and "no bounding box" as bounding box). The Hungarian matching algorithm is used to create an optimal one-to-one mapping between each of the N queries and each of the N annotations. Next, standard cross-entropy (for the classes) and a linear combination of the L1 and generalized IoU loss (for the bounding boxes) are used to optimize the parameters of the model. ## Intended uses & limitations You can use the raw model for object detection. See the [model hub](https://huggingface.co/models?search=hustvl/yolos) to look for all available YOLOS models. ### How to use Here is how to use this model: ```python from transformers import YolosFeatureExtractor, YolosForObjectDetection from PIL import Image import requests url = 'http://images.cocodataset.org/val2017/000000039769.jpg' image = Image.open(requests.get(url, stream=True).raw) feature_extractor = YolosFeatureExtractor.from_pretrained('hustvl/yolos-small') model = YolosForObjectDetection.from_pretrained('hustvl/yolos-small') inputs = feature_extractor(images=image, return_tensors="pt") outputs = model(**inputs) # model predicts bounding boxes and corresponding COCO classes logits = outputs.logits bboxes = outputs.pred_boxes ``` Currently, both the feature extractor and model support PyTorch. ## Training data The YOLOS model was pre-trained on [ImageNet-1k](https://huggingface.co/datasets/imagenet2012) and fine-tuned on [COCO 2017 object detection](https://cocodataset.org/#download), a dataset consisting of 118k/5k annotated images for training/validation respectively. ### Training The model was pre-trained for 200 epochs on ImageNet-1k and fine-tuned for 150 epochs on COCO. ## Evaluation results This model achieves an AP (average precision) of **36.1** on COCO 2017 validation. For more details regarding evaluation results, we refer to table 1 of the original paper. ### BibTeX entry and citation info ```bibtex @article{DBLP:journals/corr/abs-2106-00666, author = {Yuxin Fang and Bencheng Liao and Xinggang Wang and Jiemin Fang and Jiyang Qi and Rui Wu and Jianwei Niu and Wenyu Liu}, title = {You Only Look at One Sequence: Rethinking Transformer in Vision through Object Detection}, journal = {CoRR}, volume = {abs/2106.00666}, year = {2021}, url = {https://arxiv.org/abs/2106.00666}, eprinttype = {arXiv}, eprint = {2106.00666}, timestamp = {Fri, 29 Apr 2022 19:49:16 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-2106-00666.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ```
hustvl/yolos-base
hustvl
2024-05-08T07:48:46Z
5,096
25
transformers
[ "transformers", "pytorch", "yolos", "object-detection", "vision", "dataset:coco", "arxiv:2106.00666", "license:apache-2.0", "endpoints_compatible", "region:us" ]
object-detection
2022-04-26T09:30:39Z
--- license: apache-2.0 tags: - object-detection - vision datasets: - coco widget: - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/savanna.jpg example_title: Savanna - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/football-match.jpg example_title: Football Match - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/airport.jpg example_title: Airport --- # YOLOS (base-sized) model YOLOS model fine-tuned on COCO 2017 object detection (118k annotated images). It was introduced in the paper [You Only Look at One Sequence: Rethinking Transformer in Vision through Object Detection](https://arxiv.org/abs/2106.00666) by Fang et al. and first released in [this repository](https://github.com/hustvl/YOLOS). Disclaimer: The team releasing YOLOS did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description YOLOS is a Vision Transformer (ViT) trained using the DETR loss. Despite its simplicity, a base-sized YOLOS model is able to achieve 42 AP on COCO validation 2017 (similar to DETR and more complex frameworks such as Faster R-CNN). The model is trained using a "bipartite matching loss": one compares the predicted classes + bounding boxes of each of the N = 100 object queries to the ground truth annotations, padded up to the same length N (so if an image only contains 4 objects, 96 annotations will just have a "no object" as class and "no bounding box" as bounding box). The Hungarian matching algorithm is used to create an optimal one-to-one mapping between each of the N queries and each of the N annotations. Next, standard cross-entropy (for the classes) and a linear combination of the L1 and generalized IoU loss (for the bounding boxes) are used to optimize the parameters of the model. ## Intended uses & limitations You can use the raw model for object detection. See the [model hub](https://huggingface.co/models?search=hustvl/yolos) to look for all available YOLOS models. ### How to use Here is how to use this model: ```python from transformers import YolosFeatureExtractor, YolosForObjectDetection from PIL import Image import requests url = 'http://images.cocodataset.org/val2017/000000039769.jpg' image = Image.open(requests.get(url, stream=True).raw) feature_extractor = YolosFeatureExtractor.from_pretrained('hustvl/yolos-base') model = YolosForObjectDetection.from_pretrained('hustvl/yolos-base') inputs = feature_extractor(images=image, return_tensors="pt") outputs = model(**inputs) # model predicts bounding boxes and corresponding COCO classes logits = outputs.logits bboxes = outputs.pred_boxes ``` Currently, both the feature extractor and model support PyTorch. ## Training data The YOLOS model was pre-trained on [ImageNet-1k](https://huggingface.co/datasets/imagenet2012) and fine-tuned on [COCO 2017 object detection](https://cocodataset.org/#download), a dataset consisting of 118k/5k annotated images for training/validation respectively. ### Training The model was pre-trained for 1000 epochs on ImageNet-1k and fine-tuned for 150 epochs on COCO. ## Evaluation results This model achieves an AP (average precision) of **42.0** on COCO 2017 validation. For more details regarding evaluation results, we refer to the original paper. ### BibTeX entry and citation info ```bibtex @article{DBLP:journals/corr/abs-2106-00666, author = {Yuxin Fang and Bencheng Liao and Xinggang Wang and Jiemin Fang and Jiyang Qi and Rui Wu and Jianwei Niu and Wenyu Liu}, title = {You Only Look at One Sequence: Rethinking Transformer in Vision through Object Detection}, journal = {CoRR}, volume = {abs/2106.00666}, year = {2021}, url = {https://arxiv.org/abs/2106.00666}, eprinttype = {arXiv}, eprint = {2106.00666}, timestamp = {Fri, 29 Apr 2022 19:49:16 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-2106-00666.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ```
SenseTime/deformable-detr-single-scale
SenseTime
2024-05-08T07:47:33Z
501
0
transformers
[ "transformers", "pytorch", "safetensors", "deformable_detr", "object-detection", "vision", "dataset:coco", "arxiv:2010.04159", "license:apache-2.0", "endpoints_compatible", "region:us" ]
object-detection
2022-03-02T23:29:05Z
--- license: apache-2.0 tags: - object-detection - vision datasets: - coco widget: - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/savanna.jpg example_title: Savanna - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/football-match.jpg example_title: Football Match - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/airport.jpg example_title: Airport --- # Deformable DETR model with ResNet-50 backbone, single scale Deformable DEtection TRansformer (DETR), single scale model trained end-to-end on COCO 2017 object detection (118k annotated images). It was introduced in the paper [Deformable DETR: Deformable Transformers for End-to-End Object Detection](https://arxiv.org/abs/2010.04159) by Zhu et al. and first released in [this repository](https://github.com/fundamentalvision/Deformable-DETR). Disclaimer: The team releasing Deformable DETR did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description The DETR model is an encoder-decoder transformer with a convolutional backbone. Two heads are added on top of the decoder outputs in order to perform object detection: a linear layer for the class labels and a MLP (multi-layer perceptron) for the bounding boxes. The model uses so-called object queries to detect objects in an image. Each object query looks for a particular object in the image. For COCO, the number of object queries is set to 100. The model is trained using a "bipartite matching loss": one compares the predicted classes + bounding boxes of each of the N = 100 object queries to the ground truth annotations, padded up to the same length N (so if an image only contains 4 objects, 96 annotations will just have a "no object" as class and "no bounding box" as bounding box). The Hungarian matching algorithm is used to create an optimal one-to-one mapping between each of the N queries and each of the N annotations. Next, standard cross-entropy (for the classes) and a linear combination of the L1 and generalized IoU loss (for the bounding boxes) are used to optimize the parameters of the model. ![model image](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/deformable_detr_architecture.png) ## Intended uses & limitations You can use the raw model for object detection. See the [model hub](https://huggingface.co/models?search=sensetime/deformable-detr) to look for all available Deformable DETR models. ### How to use Here is how to use this model: ```python from transformers import AutoImageProcessor, DeformableDetrForObjectDetection import torch from PIL import Image import requests url = "http://images.cocodataset.org/val2017/000000039769.jpg" image = Image.open(requests.get(url, stream=True).raw) processor = AutoImageProcessor.from_pretrained("SenseTime/deformable-detr-single-scale") model = DeformableDetrForObjectDetection.from_pretrained("SenseTime/deformable-detr-single-scale") inputs = processor(images=image, return_tensors="pt") outputs = model(**inputs) # convert outputs (bounding boxes and class logits) to COCO API # let's only keep detections with score > 0.7 target_sizes = torch.tensor([image.size[::-1]]) results = processor.post_process_object_detection(outputs, target_sizes=target_sizes, threshold=0.7)[0] for score, label, box in zip(results["scores"], results["labels"], results["boxes"]): box = [round(i, 2) for i in box.tolist()] print( f"Detected {model.config.id2label[label.item()]} with confidence " f"{round(score.item(), 3)} at location {box}" ) ``` Currently, both the feature extractor and model support PyTorch. ## Training data The Deformable DETR model was trained on [COCO 2017 object detection](https://cocodataset.org/#download), a dataset consisting of 118k/5k annotated images for training/validation respectively. ### BibTeX entry and citation info ```bibtex @misc{https://doi.org/10.48550/arxiv.2010.04159, doi = {10.48550/ARXIV.2010.04159}, url = {https://arxiv.org/abs/2010.04159}, author = {Zhu, Xizhou and Su, Weijie and Lu, Lewei and Li, Bin and Wang, Xiaogang and Dai, Jifeng}, keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {Deformable DETR: Deformable Transformers for End-to-End Object Detection}, publisher = {arXiv}, year = {2020}, copyright = {arXiv.org perpetual, non-exclusive license} } ```
SenseTime/deformable-detr
SenseTime
2024-05-08T07:47:14Z
10,552
19
transformers
[ "transformers", "pytorch", "safetensors", "deformable_detr", "object-detection", "vision", "dataset:coco", "arxiv:2010.04159", "license:apache-2.0", "endpoints_compatible", "region:us" ]
object-detection
2022-03-02T23:29:05Z
--- license: apache-2.0 tags: - object-detection - vision datasets: - coco widget: - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/savanna.jpg example_title: Savanna - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/football-match.jpg example_title: Football Match - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/airport.jpg example_title: Airport --- # Deformable DETR model with ResNet-50 backbone Deformable DEtection TRansformer (DETR) model trained end-to-end on COCO 2017 object detection (118k annotated images). It was introduced in the paper [Deformable DETR: Deformable Transformers for End-to-End Object Detection](https://arxiv.org/abs/2010.04159) by Zhu et al. and first released in [this repository](https://github.com/fundamentalvision/Deformable-DETR). Disclaimer: The team releasing Deformable DETR did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description The DETR model is an encoder-decoder transformer with a convolutional backbone. Two heads are added on top of the decoder outputs in order to perform object detection: a linear layer for the class labels and a MLP (multi-layer perceptron) for the bounding boxes. The model uses so-called object queries to detect objects in an image. Each object query looks for a particular object in the image. For COCO, the number of object queries is set to 100. The model is trained using a "bipartite matching loss": one compares the predicted classes + bounding boxes of each of the N = 100 object queries to the ground truth annotations, padded up to the same length N (so if an image only contains 4 objects, 96 annotations will just have a "no object" as class and "no bounding box" as bounding box). The Hungarian matching algorithm is used to create an optimal one-to-one mapping between each of the N queries and each of the N annotations. Next, standard cross-entropy (for the classes) and a linear combination of the L1 and generalized IoU loss (for the bounding boxes) are used to optimize the parameters of the model. ![model image](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/deformable_detr_architecture.png) ## Intended uses & limitations You can use the raw model for object detection. See the [model hub](https://huggingface.co/models?search=sensetime/deformable-detr) to look for all available Deformable DETR models. ### How to use Here is how to use this model: ```python from transformers import AutoImageProcessor, DeformableDetrForObjectDetection import torch from PIL import Image import requests url = "http://images.cocodataset.org/val2017/000000039769.jpg" image = Image.open(requests.get(url, stream=True).raw) processor = AutoImageProcessor.from_pretrained("SenseTime/deformable-detr") model = DeformableDetrForObjectDetection.from_pretrained("SenseTime/deformable-detr") inputs = processor(images=image, return_tensors="pt") outputs = model(**inputs) # convert outputs (bounding boxes and class logits) to COCO API # let's only keep detections with score > 0.7 target_sizes = torch.tensor([image.size[::-1]]) results = processor.post_process_object_detection(outputs, target_sizes=target_sizes, threshold=0.7)[0] for score, label, box in zip(results["scores"], results["labels"], results["boxes"]): box = [round(i, 2) for i in box.tolist()] print( f"Detected {model.config.id2label[label.item()]} with confidence " f"{round(score.item(), 3)} at location {box}" ) ``` This should output: ``` Detected cat with confidence 0.856 at location [342.19, 24.3, 640.02, 372.25] Detected remote with confidence 0.739 at location [40.79, 72.78, 176.76, 117.25] Detected cat with confidence 0.859 at location [16.5, 52.84, 318.25, 470.78] ``` Currently, both the feature extractor and model support PyTorch. ## Training data The Deformable DETR model was trained on [COCO 2017 object detection](https://cocodataset.org/#download), a dataset consisting of 118k/5k annotated images for training/validation respectively. ### BibTeX entry and citation info ```bibtex @misc{https://doi.org/10.48550/arxiv.2010.04159, doi = {10.48550/ARXIV.2010.04159}, url = {https://arxiv.org/abs/2010.04159}, author = {Zhu, Xizhou and Su, Weijie and Lu, Lewei and Li, Bin and Wang, Xiaogang and Dai, Jifeng}, keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {Deformable DETR: Deformable Transformers for End-to-End Object Detection}, publisher = {arXiv}, year = {2020}, copyright = {arXiv.org perpetual, non-exclusive license} } ```
imagepipeline/cunnilingus
imagepipeline
2024-05-08T07:46:54Z
0
0
null
[ "imagepipeline", "imagepipeline.io", "text-to-image", "ultra-realistic", "license:creativeml-openrail-m", "region:us" ]
text-to-image
2024-05-08T07:46:52Z
--- license: creativeml-openrail-m tags: - imagepipeline - imagepipeline.io - text-to-image - ultra-realistic pinned: false pipeline_tag: text-to-image --- ## cunnilingus <img src="https://via.placeholder.com/468x300?text=App+Screenshot+Here" alt="Generated on Image Pipeline" style="border-radius: 10px;"> **This lora model is uploaded on [imagepipeline.io](https://imagepipeline.io/)** Model details - cunnilingus [![Try this model](https://img.shields.io/badge/try_this_model-image_pipeline-BD9319)](https://imagepipeline.io/models/cunnilingus?id=ac3be24a-bad7-45ba-8f39-5dff8c7281f8/) ## How to try this model ? You can try using it locally or send an API call to test the output quality. Get your `API_KEY` from [imagepipeline.io](https://imagepipeline.io/). No payment required. Coding in `php` `javascript` `node` etc ? Checkout our documentation [![documentation](https://img.shields.io/badge/documentation-image_pipeline-blue)](https://docs.imagepipeline.io/docs/introduction) ```python import requests import json url = "https://imagepipeline.io/sd/text2image/v1/run" payload = json.dumps({ "model_id": "sd1.5", "prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K", "negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime", "width": "512", "height": "512", "samples": "1", "num_inference_steps": "30", "safety_checker": false, "guidance_scale": 7.5, "multi_lingual": "no", "embeddings": "", "lora_models": "ac3be24a-bad7-45ba-8f39-5dff8c7281f8", "lora_weights": "0.5" }) headers = { 'Content-Type': 'application/json', 'API-Key': 'your_api_key' } response = requests.request("POST", url, headers=headers, data=payload) print(response.text) } ``` Get more ready to use `MODELS` like this for `SD 1.5` and `SDXL` : [![All models](https://img.shields.io/badge/Get%20All%20Models-image_pipeline-BD9319)](https://imagepipeline.io/models) ### API Reference #### Generate Image ```http https://api.imagepipeline.io/sd/text2image/v1 ``` | Headers | Type | Description | |:----------------------| :------- |:-------------------------------------------------------------------------------------------------------------------| | `API-Key` | `str` | Get your `API_KEY` from [imagepipeline.io](https://imagepipeline.io/) | | `Content-Type` | `str` | application/json - content type of the request body | | Parameter | Type | Description | | :-------- | :------- | :------------------------- | | `model_id` | `str` | Your base model, find available lists in [models page](https://imagepipeline.io/models) or upload your own| | `prompt` | `str` | Text Prompt. Check our [Prompt Guide](https://docs.imagepipeline.io/docs/SD-1.5/docs/extras/prompt-guide) for tips | | `num_inference_steps` | `int [1-50]` | Noise is removed with each step, resulting in a higher-quality image over time. Ideal value 30-50 (without LCM) | | `guidance_scale` | `float [1-20]` | Higher guidance scale prioritizes text prompt relevance but sacrifices image quality. Ideal value 7.5-12.5 | | `lora_models` | `str, array` | Pass the model_id(s) of LoRA models that can be found in models page | | `lora_weights` | `str, array` | Strength of the LoRA effect | --- license: creativeml-openrail-m tags: - imagepipeline - imagepipeline.io - text-to-image - ultra-realistic pinned: false pipeline_tag: text-to-image --- ### Feedback If you have any feedback, please reach out to us at hello@imagepipeline.io #### 🔗 Visit Website [![portfolio](https://img.shields.io/badge/image_pipeline-BD9319?style=for-the-badge&logo=gocd&logoColor=white)](https://imagepipeline.io/) If you are the original author of this model, please [click here](https://airtable.com/apprTaRnJbDJ8ufOx/shr4g7o9B6fWfOlUR) to add credits
itzzdeep/Mistral-7B-Instruct-v0.2-query-engine-v4-2-ckpt500-8-16-adapters
itzzdeep
2024-05-08T07:43:57Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-05-08T07:43:49Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
kknd22/RWKV6-vulkan
kknd22
2024-05-08T07:43:49Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2024-05-08T03:08:20Z
--- license: apache-2.0 ---
morten-j/pre-train_mBERTv2
morten-j
2024-05-08T07:43:09Z
123
0
transformers
[ "transformers", "tensorboard", "safetensors", "bert", "fill-mask", "generated_from_trainer", "base_model:google-bert/bert-base-multilingual-cased", "base_model:finetune:google-bert/bert-base-multilingual-cased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2024-05-07T21:28:50Z
--- license: apache-2.0 base_model: google-bert/bert-base-multilingual-cased tags: - generated_from_trainer model-index: - name: pre-train_mBERTv2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # pre-train_mBERTv2 This model is a fine-tuned version of [google-bert/bert-base-multilingual-cased](https://huggingface.co/google-bert/bert-base-multilingual-cased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.3717 - Perplexity: 3.94 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.98) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:------:|:---------------:| | 1.4925 | 1.0 | 347942 | 1.3719 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.3.0a0+ebedce2 - Datasets 2.17.1 - Tokenizers 0.15.2
gaianet/All-MiniLM-L6-v2-Embedding-GGUF
gaianet
2024-05-08T07:42:51Z
120
1
null
[ "gguf", "license:apache-2.0", "endpoints_compatible", "region:us", "feature-extraction" ]
null
2024-05-08T07:38:02Z
--- license: apache-2.0 ---
eonpark/my_awesome_model
eonpark
2024-05-08T07:42:40Z
119
0
transformers
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-07T07:37:05Z
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model-index: - name: my_awesome_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_awesome_model This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.2342 - Accuracy: 0.9314 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.234 | 1.0 | 1563 | 0.1879 | 0.9265 | | 0.1515 | 2.0 | 3126 | 0.2342 | 0.9314 | ### Framework versions - Transformers 4.30.0 - Pytorch 2.2.1+cu121 - Datasets 2.19.1 - Tokenizers 0.13.3
ShenaoZ/0.0001_sft_nodpo_3iters_bs256_5102lr_iter_2
ShenaoZ
2024-05-08T07:39:29Z
5
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "alignment-handbook", "trl", "dpo", "generated_from_trainer", "conversational", "dataset:updated", "dataset:original", "base_model:ShenaoZ/0.0001_sft_nodpo_3iters_bs256_5102lr_iter_1", "base_model:finetune:ShenaoZ/0.0001_sft_nodpo_3iters_bs256_5102lr_iter_1", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T06:37:43Z
--- license: mit base_model: ShenaoZ/0.0001_sft_nodpo_3iters_bs256_5102lr_iter_1 tags: - alignment-handbook - trl - dpo - generated_from_trainer - trl - dpo - generated_from_trainer datasets: - updated - original model-index: - name: 0.0001_sft_nodpo_3iters_bs256_5102lr_iter_2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # 0.0001_sft_nodpo_3iters_bs256_5102lr_iter_2 This model is a fine-tuned version of [ShenaoZ/0.0001_sft_nodpo_3iters_bs256_5102lr_iter_1](https://huggingface.co/ShenaoZ/0.0001_sft_nodpo_3iters_bs256_5102lr_iter_1) on the updated and the original datasets. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-07 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - gradient_accumulation_steps: 4 - total_train_batch_size: 256 - total_eval_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 1 ### Training results ### Framework versions - Transformers 4.40.2 - Pytorch 2.1.2+cu121 - Datasets 2.14.6 - Tokenizers 0.19.1
shanmukeshwar/oc-finetuned
shanmukeshwar
2024-05-08T07:32:07Z
0
0
transformers
[ "transformers", "safetensors", "autotrain", "text-generation-inference", "text-generation", "peft", "conversational", "license:other", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T07:20:06Z
--- tags: - autotrain - text-generation-inference - text-generation - peft library_name: transformers widget: - messages: - role: user content: What is your favorite condiment? license: other --- # Model Trained Using AutoTrain This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain). # Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_path = "PATH_TO_THIS_REPO" tokenizer = AutoTokenizer.from_pretrained(model_path) model = AutoModelForCausalLM.from_pretrained( model_path, device_map="auto", torch_dtype='auto' ).eval() # Prompt content: "hi" messages = [ {"role": "user", "content": "hi"} ] input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt') output_ids = model.generate(input_ids.to('cuda')) response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True) # Model response: "Hello! How can I assist you today?" print(response) ```
novalentino94/my_awesome_qa_model
novalentino94
2024-05-08T07:27:45Z
107
0
transformers
[ "transformers", "tensorboard", "safetensors", "distilbert", "question-answering", "generated_from_trainer", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:apache-2.0", "endpoints_compatible", "region:us" ]
question-answering
2024-05-02T03:33:43Z
--- license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer model-index: - name: my_awesome_qa_model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # my_awesome_qa_model This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.8364 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | No log | 1.0 | 250 | 2.4768 | | 2.7735 | 2.0 | 500 | 1.9128 | | 2.7735 | 3.0 | 750 | 1.8364 | ### Framework versions - Transformers 4.40.1 - Pytorch 2.2.1+cu121 - Datasets 2.19.0 - Tokenizers 0.19.1
abc88767/model100
abc88767
2024-05-08T07:27:15Z
132
0
transformers
[ "transformers", "safetensors", "stablelm", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T07:25:39Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
four-two-labs/phi3-nord-10k
four-two-labs
2024-05-08T07:25:41Z
4
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:microsoft/Phi-3-mini-4k-instruct", "base_model:adapter:microsoft/Phi-3-mini-4k-instruct", "region:us" ]
null
2024-05-08T07:25:30Z
--- library_name: peft base_model: microsoft/Phi-3-mini-4k-instruct --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.10.0
anezatra/gpt2-nos-124M
anezatra
2024-05-08T07:24:50Z
164
0
transformers
[ "transformers", "tensorboard", "safetensors", "gpt2", "text-generation", "en", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-03T18:40:29Z
--- language: - en pipeline_tag: text-generation --- ## GPT-2 NOS INSTAGRAM ANALYZER AI ![examples](https://huggingface.co/anezatra/gpt2-nos-124M/raw/main/banner.jpg) Nos AI is an artificial intelligence software that utilizes the GPT-2 language model to analyze Instagram users in real-time at specified intervals and communicates this information to the user. For instance, Nos attempts to find out whom an Instagram user most recently sent a request to or where a specific account's name has appeared, either in posts or biographies, by examining the accounts followed by the target user. All these operations are conducted by leveraging Instagram APIs. To use the program, Nos needs to log in to any account. ## Model Description The GPT-2 model used by Nos was trained on a small dataset consisting of specific actions. The current GPT-2 Model has 124 million hyperparameters. It generates text based on users' actions on their accounts. Nos GPT2 first collects the accounts followed by the target account. During the initial collection, it perceives them as newly followed accounts and records the initially checked users. Later on, it checks whether the analyzed user's name appears in the accounts they follow. For this purpose, it focuses on the writings, comments, and biographies of the accounts followed by the analyzed user. If the analyzed user's name is found among the followed accounts, Nos promptly informs you. Subsequently, it attempts to collect the open-source data of the analyzed user from the internet, completing the initial check this way. In the next analysis, if the analyzed user has followed someone new within the specified time frame, Nos directly informs you about this person and analyzes all their information. - Check out the github page for the [program](https://github.com/anezatra/nos-ai) ## Training - This GPT-2 model is rated for an average of 1 hour with an A100 GPU. **Authors** - **Developed by:** Anezatra - **Model type:** GPT2 - **Contacts:** https://github.com/anezatra
aaron-di/YamshadowExperiment28-7B-DareLinear
aaron-di
2024-05-08T07:24:22Z
7
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "automerger", "base_model:automerger/YamShadow-7B", "base_model:merge:automerger/YamShadow-7B", "base_model:yam-peleg/Experiment28-7B", "base_model:merge:yam-peleg/Experiment28-7B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T07:21:09Z
--- license: apache-2.0 tags: - merge - mergekit - lazymergekit - automerger base_model: - automerger/YamShadow-7B - yam-peleg/Experiment28-7B --- ## 🧩 Configuration ```yaml models: - model: automerger/YamShadow-7B parameters: density: 0.5 weight: 0.5 - model: yam-peleg/Experiment28-7B parameters: density: 0.5 weight: 0.5 merge_method: dare_linear base_model: automerger/YamShadow-7B dtype: float16 random_seed: 0 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "aaron-di/YamshadowExperiment28-7B-DareLinear" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
imagepipeline/dyer
imagepipeline
2024-05-08T07:21:23Z
0
0
null
[ "imagepipeline", "imagepipeline.io", "text-to-image", "ultra-realistic", "license:creativeml-openrail-m", "region:us" ]
text-to-image
2024-05-08T07:21:20Z
--- license: creativeml-openrail-m tags: - imagepipeline - imagepipeline.io - text-to-image - ultra-realistic pinned: false pipeline_tag: text-to-image --- ## dyer <img src="https://via.placeholder.com/468x300?text=App+Screenshot+Here" alt="Generated on Image Pipeline" style="border-radius: 10px;"> **This lora model is uploaded on [imagepipeline.io](https://imagepipeline.io/)** Model details - dyer [![Try this model](https://img.shields.io/badge/try_this_model-image_pipeline-BD9319)](https://imagepipeline.io/models/dyer?id=0c4dfd9b-8103-452c-94a4-bee84eca17fd/) ## How to try this model ? You can try using it locally or send an API call to test the output quality. Get your `API_KEY` from [imagepipeline.io](https://imagepipeline.io/). No payment required. Coding in `php` `javascript` `node` etc ? Checkout our documentation [![documentation](https://img.shields.io/badge/documentation-image_pipeline-blue)](https://docs.imagepipeline.io/docs/introduction) ```python import requests import json url = "https://imagepipeline.io/sd/text2image/v1/run" payload = json.dumps({ "model_id": "sd1.5", "prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K", "negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime", "width": "512", "height": "512", "samples": "1", "num_inference_steps": "30", "safety_checker": false, "guidance_scale": 7.5, "multi_lingual": "no", "embeddings": "", "lora_models": "0c4dfd9b-8103-452c-94a4-bee84eca17fd", "lora_weights": "0.5" }) headers = { 'Content-Type': 'application/json', 'API-Key': 'your_api_key' } response = requests.request("POST", url, headers=headers, data=payload) print(response.text) } ``` Get more ready to use `MODELS` like this for `SD 1.5` and `SDXL` : [![All models](https://img.shields.io/badge/Get%20All%20Models-image_pipeline-BD9319)](https://imagepipeline.io/models) ### API Reference #### Generate Image ```http https://api.imagepipeline.io/sd/text2image/v1 ``` | Headers | Type | Description | |:----------------------| :------- |:-------------------------------------------------------------------------------------------------------------------| | `API-Key` | `str` | Get your `API_KEY` from [imagepipeline.io](https://imagepipeline.io/) | | `Content-Type` | `str` | application/json - content type of the request body | | Parameter | Type | Description | | :-------- | :------- | :------------------------- | | `model_id` | `str` | Your base model, find available lists in [models page](https://imagepipeline.io/models) or upload your own| | `prompt` | `str` | Text Prompt. Check our [Prompt Guide](https://docs.imagepipeline.io/docs/SD-1.5/docs/extras/prompt-guide) for tips | | `num_inference_steps` | `int [1-50]` | Noise is removed with each step, resulting in a higher-quality image over time. Ideal value 30-50 (without LCM) | | `guidance_scale` | `float [1-20]` | Higher guidance scale prioritizes text prompt relevance but sacrifices image quality. Ideal value 7.5-12.5 | | `lora_models` | `str, array` | Pass the model_id(s) of LoRA models that can be found in models page | | `lora_weights` | `str, array` | Strength of the LoRA effect | --- license: creativeml-openrail-m tags: - imagepipeline - imagepipeline.io - text-to-image - ultra-realistic pinned: false pipeline_tag: text-to-image --- ### Feedback If you have any feedback, please reach out to us at hello@imagepipeline.io #### 🔗 Visit Website [![portfolio](https://img.shields.io/badge/image_pipeline-BD9319?style=for-the-badge&logo=gocd&logoColor=white)](https://imagepipeline.io/) If you are the original author of this model, please [click here](https://airtable.com/apprTaRnJbDJ8ufOx/shr4g7o9B6fWfOlUR) to add credits
ZahraRahimiii/q-FrozenLake-v1-4x4-Slippery
ZahraRahimiii
2024-05-08T07:21:11Z
0
0
null
[ "FrozenLake-v1-4x4", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
reinforcement-learning
2024-05-08T07:21:08Z
--- tags: - FrozenLake-v1-4x4 - q-learning - reinforcement-learning - custom-implementation model-index: - name: q-FrozenLake-v1-4x4-Slippery results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: FrozenLake-v1-4x4 type: FrozenLake-v1-4x4 metrics: - type: mean_reward value: 0.47 +/- 0.50 name: mean_reward verified: false --- # **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="ZahraRahimiii/q-FrozenLake-v1-4x4-Slippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
Edgar404/donut
Edgar404
2024-05-08T07:16:09Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2024-05-08T07:16:07Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
ShenaoZ/0.0001_sft_nodpo_5iters_bs256_5102lr_iter_2
ShenaoZ
2024-05-08T07:11:13Z
5
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "alignment-handbook", "trl", "dpo", "generated_from_trainer", "conversational", "dataset:updated", "dataset:original", "base_model:ShenaoZ/0.0001_sft_nodpo_5iters_bs256_5102lr_iter_1", "base_model:finetune:ShenaoZ/0.0001_sft_nodpo_5iters_bs256_5102lr_iter_1", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T06:31:35Z
--- license: mit base_model: ShenaoZ/0.0001_sft_nodpo_5iters_bs256_5102lr_iter_1 tags: - alignment-handbook - trl - dpo - generated_from_trainer - trl - dpo - generated_from_trainer datasets: - updated - original model-index: - name: 0.0001_sft_nodpo_5iters_bs256_5102lr_iter_2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # 0.0001_sft_nodpo_5iters_bs256_5102lr_iter_2 This model is a fine-tuned version of [ShenaoZ/0.0001_sft_nodpo_5iters_bs256_5102lr_iter_1](https://huggingface.co/ShenaoZ/0.0001_sft_nodpo_5iters_bs256_5102lr_iter_1) on the updated and the original datasets. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-07 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - gradient_accumulation_steps: 4 - total_train_batch_size: 256 - total_eval_batch_size: 64 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 1 ### Training results ### Framework versions - Transformers 4.40.2 - Pytorch 2.1.2+cu121 - Datasets 2.14.6 - Tokenizers 0.19.1
knosing/japanese_ner_model
knosing
2024-05-08T07:06:22Z
191
0
transformers
[ "transformers", "pytorch", "safetensors", "bert", "token-classification", "ner", "named entity recognition", "stockmark ner", "japanese named entity recognition", "japanese ner", "ja", "en", "dataset:stockmark/ner-wikipedia-dataset", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
token-classification
2024-05-08T06:15:37Z
--- license: apache-2.0 datasets: - stockmark/ner-wikipedia-dataset language: - ja - en metrics: - f1 - recall - precision - accuracy library_name: transformers pipeline_tag: token-classification tags: - ner - named entity recognition - stockmark ner - bert - japanese named entity recognition - japanese ner - transformers --- ### Model Description This model is a fine-tuned version of the `tohoku-nlp/bert-base-japanese-v3`, specifically optimized for Named Entity Recognition (NER) tasks. It is fine-tuned using a Japanese named entity extraction dataset derived from Wikipedia, which was developed and made publicly available by Stockmark Inc. ([NER Wikipedia Dataset](https://github.com/stockmarkteam/ner-wikipedia-dataset)). ### Intended Use This model is intended for use in tasks that require the identification and categorization of named entities within Japanese text. It is suitable for various applications in natural language processing where understanding the specific names of people, organizations, locations, etc., is crucial. ### How to Use You can use this model for NER tasks with the following simple code snippet: ```python from transformers import AutoModelForTokenClassification, AutoTokenizer import torch model_name = "knosing/japanese_ner_model" tokenizer = AutoTokenizer.from_pretrained("tohoku-nlp/bert-base-japanese-v3") model = AutoModelForTokenClassification.from_pretrained(model_name) ``` ### Model Performance The model has been evaluated on various entity types to assess its precision, recall, F1 score, and overall accuracy. Below is the detailed performance breakdown by entity type: #### Overall Metrics - **Overall Precision:** 0.8379 - **Overall Recall:** 0.8477 - **Overall F1 Score:** 0.8428 - **Overall Accuracy:** 0.9684 #### Performance by Entity Type - **Other Organization Names (`の他の組織名`):** - **Precision:** 0.71875 - **Recall:** 0.69 - **F1 Score:** 0.7041 - **Sample Count:** 100 - **Event Names (`ベント名`):** - **Precision:** 0.85 - **Recall:** 0.8586 - **F1 Score:** 0.8543 - **Sample Count:** 99 - **Personal Names (`人名`):** - **Precision:** 0.8171 - **Recall:** 0.8664 - **F1 Score:** 0.8410 - **Sample Count:** 232 - **Generic Names (`名`):** - **Precision:** 0.8986 - **Recall:** 0.9376 - **F1 Score:** 0.9177 - **Sample Count:** 529 - **Product Names (`品名`):** - **Precision:** 0.6522 - **Recall:** 0.5906 - **F1 Score:** 0.6198 - **Sample Count:** 127 - **Government Organization Names (`治的組織名`):** - **Precision:** 0.9160 - **Recall:** 0.8276 - **F1 Score:** 0.8696 - **Sample Count:** 145 - **Facility Names (`設名`):** - **Precision:** 0.7905 - **Recall:** 0.8357 - **F1 Score:** 0.8125 - **Sample Count:** 140 ### Note You might not able to use the model with huggingface Inference API. The intended use for the model is given in the following repository: [KeshavSingh29/fa_ner_japanese](https://github.com/KeshavSingh29/fa_ner_japanese) If you have any questions, please feel free to contact me or raise an issue at the above repo.
Defetya/ru-adapt-gemma-7b
Defetya
2024-05-08T06:57:15Z
6
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-08T06:52:57Z
--- license: apache-2.0 ---
mradermacher/L3-Picaro-8B-GGUF
mradermacher
2024-05-08T06:56:56Z
52
1
transformers
[ "transformers", "gguf", "en", "base_model:royallab/L3-Picaro-8B", "base_model:quantized:royallab/L3-Picaro-8B", "endpoints_compatible", "region:us" ]
null
2024-05-08T04:44:07Z
--- base_model: royallab/L3-Picaro-8B language: - en library_name: transformers quantized_by: mradermacher --- ## About <!-- ### quantize_version: 2 --> <!-- ### output_tensor_quantised: 1 --> <!-- ### convert_type: hf --> <!-- ### vocab_type: --> static quants of https://huggingface.co/royallab/L3-Picaro-8B <!-- provided-files --> weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. ## Usage If you are unsure how to use GGUF files, refer to one of [TheBloke's READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for more details, including on how to concatenate multi-part files. ## Provided Quants (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.Q2_K.gguf) | Q2_K | 3.3 | | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.IQ3_XS.gguf) | IQ3_XS | 3.6 | | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.Q3_K_S.gguf) | Q3_K_S | 3.8 | | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.IQ3_S.gguf) | IQ3_S | 3.8 | beats Q3_K* | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.IQ3_M.gguf) | IQ3_M | 3.9 | | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.Q3_K_M.gguf) | Q3_K_M | 4.1 | lower quality | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.Q3_K_L.gguf) | Q3_K_L | 4.4 | | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.IQ4_XS.gguf) | IQ4_XS | 4.6 | | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.Q4_K_S.gguf) | Q4_K_S | 4.8 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.Q4_K_M.gguf) | Q4_K_M | 5.0 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.Q5_K_S.gguf) | Q5_K_S | 5.7 | | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.Q5_K_M.gguf) | Q5_K_M | 5.8 | | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.Q6_K.gguf) | Q6_K | 6.7 | very good quality | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.Q8_0.gguf) | Q8_0 | 8.6 | fast, best quality | | [GGUF](https://huggingface.co/mradermacher/L3-Picaro-8B-GGUF/resolve/main/L3-Picaro-8B.f16.gguf) | f16 | 16.2 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 ## FAQ / Model Request See https://huggingface.co/mradermacher/model_requests for some answers to questions you might have and/or if you want some other model quantized. ## Thanks I thank my company, [nethype GmbH](https://www.nethype.de/), for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. <!-- end -->
mradermacher/l3-badger-mushroom-4x8b-GGUF
mradermacher
2024-05-08T06:55:46Z
43
1
transformers
[ "transformers", "gguf", "llama-3", "en", "base_model:maldv/l3-badger-mushroom-4x8b", "base_model:quantized:maldv/l3-badger-mushroom-4x8b", "license:cc-by-nc-4.0", "endpoints_compatible", "region:us", "conversational" ]
null
2024-05-08T05:26:07Z
--- base_model: maldv/l3-badger-mushroom-4x8b language: - en library_name: transformers license: cc-by-nc-4.0 quantized_by: mradermacher tags: - llama-3 --- ## About <!-- ### quantize_version: 2 --> <!-- ### output_tensor_quantised: 1 --> <!-- ### convert_type: hfhfix --> <!-- ### vocab_type: --> static quants of https://huggingface.co/maldv/l3-badger-mushroom-4x8b <!-- provided-files --> weighted/imatrix quants are available at https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-i1-GGUF ## Usage If you are unsure how to use GGUF files, refer to one of [TheBloke's READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for more details, including on how to concatenate multi-part files. ## Provided Quants (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.Q2_K.gguf) | Q2_K | 9.4 | | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.IQ3_XS.gguf) | IQ3_XS | 10.5 | | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.Q3_K_S.gguf) | Q3_K_S | 11.0 | | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.IQ3_S.gguf) | IQ3_S | 11.1 | beats Q3_K* | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.IQ3_M.gguf) | IQ3_M | 11.2 | | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.Q3_K_M.gguf) | Q3_K_M | 12.2 | lower quality | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.Q3_K_L.gguf) | Q3_K_L | 13.1 | | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.IQ4_XS.gguf) | IQ4_XS | 13.7 | | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.Q4_K_S.gguf) | Q4_K_S | 14.4 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.Q4_K_M.gguf) | Q4_K_M | 15.3 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.Q5_K_S.gguf) | Q5_K_S | 17.3 | | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.Q5_K_M.gguf) | Q5_K_M | 17.8 | | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.Q6_K.gguf) | Q6_K | 20.6 | very good quality | | [GGUF](https://huggingface.co/mradermacher/l3-badger-mushroom-4x8b-GGUF/resolve/main/l3-badger-mushroom-4x8b.Q8_0.gguf) | Q8_0 | 26.6 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 ## FAQ / Model Request See https://huggingface.co/mradermacher/model_requests for some answers to questions you might have and/or if you want some other model quantized. ## Thanks I thank my company, [nethype GmbH](https://www.nethype.de/), for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. <!-- end -->
JaikiratSinghSeth/vit-base-patch16-224-finetuned-flower
JaikiratSinghSeth
2024-05-08T06:55:02Z
166
0
transformers
[ "transformers", "pytorch", "vit", "image-classification", "generated_from_trainer", "dataset:imagefolder", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2024-05-08T06:44:27Z
--- license: apache-2.0 tags: - generated_from_trainer datasets: - imagefolder model-index: - name: vit-base-patch16-224-finetuned-flower results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # vit-base-patch16-224-finetuned-flower This model is a fine-tuned version of [google/vit-base-patch16-224](https://huggingface.co/google/vit-base-patch16-224) on the imagefolder dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results ### Framework versions - Transformers 4.24.0 - Pytorch 2.2.1+cu121 - Datasets 2.7.1 - Tokenizers 0.13.3
opennyaiorg/InExtractiveSummarizer
opennyaiorg
2024-05-08T06:49:59Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2024-05-08T06:49:11Z
--- license: apache-2.0 ---
hobee/CommentLM-7B
hobee
2024-05-08T06:49:45Z
2
1
transformers
[ "transformers", "pytorch", "internlm2", "feature-extraction", "custom_code", "license:other", "region:us" ]
feature-extraction
2024-05-08T03:50:50Z
--- license: other license_name: other license_link: LICENSE ---
DUAL-GPO/zephyr-7b-gpo-v8-i1
DUAL-GPO
2024-05-08T06:41:39Z
1
0
peft
[ "peft", "tensorboard", "safetensors", "mistral", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "dataset:HuggingFaceH4/ultrafeedback_binarized", "base_model:DUAL-GPO/zephyr-7b-gpo-final-i0", "base_model:adapter:DUAL-GPO/zephyr-7b-gpo-final-i0", "license:mit", "region:us" ]
null
2024-05-07T20:59:21Z
--- license: mit library_name: peft tags: - alignment-handbook - generated_from_trainer - trl - dpo - generated_from_trainer base_model: DUAL-GPO/zephyr-7b-gpo-final-i0 datasets: - HuggingFaceH4/ultrafeedback_binarized model-index: - name: zephyr-7b-gpo-v8-i1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # zephyr-7b-gpo-v8-i1 This model is a fine-tuned version of [DUAL-GPO/zephyr-7b-gpo-final-i0](https://huggingface.co/DUAL-GPO/zephyr-7b-gpo-final-i0) on the HuggingFaceH4/ultrafeedback_binarized dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - distributed_type: multi-GPU - gradient_accumulation_steps: 2 - total_train_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 1 ### Training results ### Framework versions - PEFT 0.7.1 - Transformers 4.36.2 - Pytorch 2.1.2+cu121 - Datasets 2.14.6 - Tokenizers 0.15.2
misaza/Sentimiento-appmovilesXLM
misaza
2024-05-08T06:37:23Z
118
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:cardiffnlp/twitter-xlm-roberta-base-sentiment", "base_model:finetune:cardiffnlp/twitter-xlm-roberta-base-sentiment", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-08T02:49:38Z
--- base_model: cardiffnlp/twitter-xlm-roberta-base-sentiment tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: Sentimiento-appmovilesXLM results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Sentimiento-appmovilesXLM This model is a fine-tuned version of [cardiffnlp/twitter-xlm-roberta-base-sentiment](https://huggingface.co/cardiffnlp/twitter-xlm-roberta-base-sentiment) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.2940 - Accuracy: 0.9337 - F1: 0.8559 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | No log | 1.0 | 332 | 0.2940 | 0.9337 | 0.8559 | | 0.3386 | 2.0 | 664 | 0.3005 | 0.9307 | 0.8290 | ### Framework versions - Transformers 4.40.1 - Pytorch 2.2.1+cu121 - Datasets 2.19.1 - Tokenizers 0.19.1
CyrexPro/pegasus-xsum-finetuned-cnn_dailymail
CyrexPro
2024-05-08T06:35:38Z
13
0
transformers
[ "transformers", "tensorboard", "safetensors", "pegasus", "text2text-generation", "generated_from_trainer", "base_model:google/pegasus-xsum", "base_model:finetune:google/pegasus-xsum", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2023-07-25T14:01:35Z
--- base_model: google/pegasus-xsum tags: - generated_from_trainer metrics: - rouge model-index: - name: pegasus-xsum-finetuned-cnn_dailymail results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # pegasus-xsum-finetuned-cnn_dailymail This model is a fine-tuned version of [google/pegasus-xsum](https://huggingface.co/google/pegasus-xsum) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.8958 - Rouge1: 45.7795 - Rouge2: 23.3182 - Rougel: 32.9241 - Rougelsum: 42.3126 - Bleu 1: 35.4715 - Bleu 2: 24.0726 - Bleu 3: 17.9591 - Meteor: 32.8897 - Lungime rezumat: 43.3773 - Lungime original: 48.6937 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5.6e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Bleu 1 | Bleu 2 | Bleu 3 | Meteor | Lungime rezumat | Lungime original | |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|:-------:|:-------:|:-------:|:---------------:|:----------------:| | 1.1281 | 1.0 | 14330 | 0.9373 | 44.64 | 22.2111 | 32.0228 | 41.1223 | 34.4946 | 23.079 | 17.0673 | 31.8685 | 43.543 | 48.6937 | | 0.9091 | 2.0 | 28660 | 0.9095 | 45.0713 | 22.7428 | 32.4247 | 41.554 | 34.9397 | 23.5631 | 17.5094 | 32.1814 | 43.3467 | 48.6937 | | 0.8455 | 3.0 | 42990 | 0.8982 | 45.5457 | 23.1315 | 32.7153 | 42.0349 | 35.2659 | 23.8773 | 17.8174 | 32.7185 | 43.5743 | 48.6937 | | 0.8076 | 4.0 | 57320 | 0.8958 | 45.7795 | 23.3182 | 32.9241 | 42.3126 | 35.4715 | 24.0726 | 17.9591 | 32.8897 | 43.3773 | 48.6937 | ### Framework versions - Transformers 4.40.0 - Pytorch 2.2.2+cu118 - Datasets 2.19.0 - Tokenizers 0.19.1
rj1ALINT/day-time
rj1ALINT
2024-05-08T06:24:58Z
29
0
diffusers
[ "diffusers", "safetensors", "text-to-image", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-05-08T06:23:52Z
--- license: creativeml-openrail-m tags: - text-to-image --- ### day_time on Stable Diffusion via Dreambooth #### model by rj1ALINT This your the Stable Diffusion model fine-tuned the day_time concept taught to Stable Diffusion with Dreambooth. It can be used by modifying the `instance_prompt`: **<dashcam footage > of a car driving at day time** You can also train your own concepts and upload them to the library by using [this notebook](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_dreambooth_training.ipynb). And you can run your new concept via `diffusers`: [Colab Notebook for Inference](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_dreambooth_inference.ipynb), [Spaces with the Public Concepts loaded](https://huggingface.co/spaces/sd-dreambooth-library/stable-diffusion-dreambooth-concepts) Here are the images used for training this concept: ![image 0](https://huggingface.co/rj1ALINT/day-time/resolve/main/concept_images/3.jpeg) ![image 1](https://huggingface.co/rj1ALINT/day-time/resolve/main/concept_images/0.jpeg) ![image 2](https://huggingface.co/rj1ALINT/day-time/resolve/main/concept_images/2.jpeg) ![image 3](https://huggingface.co/rj1ALINT/day-time/resolve/main/concept_images/4.jpeg) ![image 4](https://huggingface.co/rj1ALINT/day-time/resolve/main/concept_images/1.jpeg)
Irfanarshad/Cifar10_model
Irfanarshad
2024-05-08T06:20:47Z
0
0
null
[ "code", "image-classification", "en", "dataset:cifar10", "doi:10.57967/hf/2211", "license:unknown", "region:us" ]
image-classification
2024-05-08T06:04:24Z
--- license: unknown datasets: - cifar10 language: - en metrics: - accuracy pipeline_tag: image-classification tags: - code --- # Model Card for Model ID # Convolutional Neural Network (CNN) for CIFAR-10 Dataset This repository contains a Convolutional Neural Network (CNN) model trained on the CIFAR-10 dataset. CIFAR-10 is a popular benchmark dataset in the field of computer vision, consisting of 60,000 32x32 color images in 10 classes, with 6,000 images per class. # Model Architecture: The CNN model architecture consists of several convolutional layers followed by max-pooling layers, batch normalization, and ReLU activation functions. The final layers typically include one or more fully connected layers followed by a softmax layer for classification. # Dataset: CIFAR-10 dataset is used for training and evaluation. It comprises 50,000 training images and 10,000 test images across 10 classes, including airplanes, cars, birds, cats, deer, dogs, frogs, horses, ships, and trucks. # Training: The model is trained using stochastic gradient descent (SGD) or other optimization algorithms with appropriate hyperparameters such as learning rate, batch size, and number of epochs. # Evaluation: The performance of the model is evaluated using metrics such as accuracy, precision, recall, and F1-score on a separate validation set or through cross-validation. # Usage: The trained model can be used for various computer vision tasks, including image classification, object detection, and image segmentation. - **Developed by:** Irfan Arshad - **Funded by [optional]:** PAF-IAST - **Shared by [optional]:** Irfan Arshad - **Model type:** Classification Model - **Language(s) (NLP):** Python Tensorflow Deep Learning ### Results ![image/png](https://cdn-uploads.huggingface.co/production/uploads/663b152c699810ce7e540fe7/ScWiPc6-xnOraO6aZwnal.png) #### Software Google Colab Visual Studio Code Web Browser
stablediffusionapi/realisticmixmiracle-new
stablediffusionapi
2024-05-08T06:17:56Z
29
1
diffusers
[ "diffusers", "safetensors", "modelslab.com", "stable-diffusion-api", "text-to-image", "ultra-realistic", "license:creativeml-openrail-m", "autotrain_compatible", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
text-to-image
2024-04-16T08:58:39Z
--- license: creativeml-openrail-m tags: - modelslab.com - stable-diffusion-api - text-to-image - ultra-realistic pinned: true --- # API Inference ![generated from modelslab.com](https://cdn2.stablediffusionapi.com/generations/bf190b5a-fe19-437c-ba05-82f29cb1f7ad-0.png) ## Get API Key Get API key from [ModelsLab API](http://modelslab.com), No Payment needed. Replace Key in below code, change **model_id** to "realisticmixmiracle-new" Coding in PHP/Node/Java etc? Have a look at docs for more code examples: [View docs](https://modelslab.com/docs) Try model for free: [Generate Images](https://modelslab.com/models/realisticmixmiracle-new) Model link: [View model](https://modelslab.com/models/realisticmixmiracle-new) View all models: [View Models](https://modelslab.com/models) import requests import json url = "https://modelslab.com/api/v6/images/text2img" payload = json.dumps({ "key": "your_api_key", "model_id": "realisticmixmiracle-new", "prompt": "ultra realistic close up portrait ((beautiful pale cyberpunk female with heavy black eyeliner)), blue eyes, shaved side haircut, hyper detail, cinematic lighting, magic neon, dark red city, Canon EOS R3, nikon, f/1.4, ISO 200, 1/160s, 8K, RAW, unedited, symmetrical balance, in-frame, 8K", "negative_prompt": "painting, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, deformed, ugly, blurry, bad anatomy, bad proportions, extra limbs, cloned face, skinny, glitchy, double torso, extra arms, extra hands, mangled fingers, missing lips, ugly face, distorted face, extra legs, anime", "width": "512", "height": "512", "samples": "1", "num_inference_steps": "30", "safety_checker": "no", "enhance_prompt": "yes", "seed": None, "guidance_scale": 7.5, "multi_lingual": "no", "panorama": "no", "self_attention": "no", "upscale": "no", "embeddings": "embeddings_model_id", "lora": "lora_model_id", "webhook": None, "track_id": None }) headers = { 'Content-Type': 'application/json' } response = requests.request("POST", url, headers=headers, data=payload) print(response.text) > Use this coupon code to get 25% off **DMGG0RBN**
olivernoah/OSTtoPSTAPP-Outlook-PST-password-recovery-software
olivernoah
2024-05-08T06:03:07Z
0
0
null
[ "region:us" ]
null
2024-05-08T06:02:03Z
Users can Recover Outlook PST Password with the help of OSTtoPSTAPP Outlook PST password recovery software. Any type of PST file password can be recovered with this program. Outlook PST Password Recovery Software is user-friendly, anyone can recover and reset Outlook PST password. Users can access the password for several PST files of Outlook with the software's advanced feature. Users don't have problems removing the password from any secret PST file and can load PST folders independently. The software can be used for recovering the password for any version of Microsoft Outlook. The software-supported PST file password for Outlook 2021, 2019, 2016, 2013, 2010, 2007, 2003, and others is processed. Both ANSI and Unicode PST files are supported by it properly. Even users can use any editions of Windows 11, 10, 8, 8.1, 7, XP, or Vista with this software. The program is free to download and use for Users. Read More:- https://www.osttopstapp.com/pst-password-recovery.html
LocalDoc/sentiment_analysis_azerbaijani
LocalDoc
2024-05-08T06:02:58Z
112
1
transformers
[ "transformers", "safetensors", "xlm-roberta", "text-classification", "sentiment", "analysis", "azerbaijani", "az", "dataset:LocalDoc/sentiments_dataset_azerbaijani", "doi:10.57967/hf/2361", "license:cc-by-nc-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-07T18:17:05Z
--- license: cc-by-nc-4.0 language: - az pipeline_tag: text-classification tags: - sentiment - analysis - azerbaijani widget: - text: Bu mənim xoşuma gəlir datasets: - LocalDoc/sentiments_dataset_azerbaijani --- # Sentiment Analysis Model for Azerbaijani Text This repository hosts a fine-tuned XLM-RoBERTa model for sentiment analysis on Azerbaijani text. The model is capable of classifying text into three categories: negative, neutral, and positive. ## Model Description The model is based on `xlm-roberta-base`, which has been fine-tuned on a diverse dataset of Azerbaijani text samples. It is designed to understand the sentiment expressed in texts and classify them accordingly. ## How to Use You can use this model directly with a pipeline for text classification, or you can use it with the `transformers` library for more custom usage, as shown in the example below. ### Quick Start First, install the transformers library if you haven't already: ```bash pip install transformers ``` ```python from transformers import AutoModelForSequenceClassification, XLMRobertaTokenizer import torch # Load the model and tokenizer from Hugging Face Hub model_name = "LocalDoc/sentiment_analysis_azerbaijani" tokenizer = XLMRobertaTokenizer.from_pretrained(model_name) model = AutoModelForSequenceClassification.from_pretrained(model_name) def predict_sentiment(text): # Encode the text using the tokenizer inputs = tokenizer(text, return_tensors="pt", padding=True, truncation=True, max_length=128) # Get predictions from the model with torch.no_grad(): outputs = model(**inputs) # Convert logits to probabilities using softmax probs = torch.nn.functional.softmax(outputs.logits, dim=-1) # Get the highest probability and corresponding label top_prob, top_label = torch.max(probs, dim=-1) labels = ["negative", "neutral", "positive"] # Return the label with the highest probability return labels[top_label], top_prob # Example text text = "Bu mənim xoşuma gəlir" # Get the sentiment predicted_label, probability = predict_sentiment(text) print(f"Predicted sentiment: {predicted_label} with a probability of {probability.item():.4f}") ``` ## Sentiment Label Information The model outputs a label for each prediction, corresponding to one of the sentiment categories listed below. Each label is associated with a specific sentiment as detailed in the following table: | Label | Sentiment | |-------|-----------| | 0 | Negative | | 1 | Neutral | | 2 | Positive | License The dataset is licensed under the Creative Commons Attribution-NonCommercial 4.0 International license. This license allows you to freely share and redistribute the dataset with attribution to the source but prohibits commercial use and the creation of derivative works. Contact information If you have any questions or suggestions, please contact us at [v.resad.89@gmail.com].
AmirlyPhd/final_V2-bert-after-adding-new-words-text-classification-model
AmirlyPhd
2024-05-08T06:02:49Z
110
0
transformers
[ "transformers", "safetensors", "bert", "text-classification", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-05-08T06:02:29Z
--- tags: - generated_from_trainer metrics: - accuracy - f1 - precision - recall model-index: - name: final_V2-bert-after-adding-new-words-text-classification-model results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # final_V2-bert-after-adding-new-words-text-classification-model This model was trained from scratch on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.1494 - Accuracy: 0.9716 - F1: 0.8348 - Precision: 0.8317 - Recall: 0.8385 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 3 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:| | 1.8136 | 0.11 | 50 | 1.7501 | 0.3470 | 0.1733 | 0.3034 | 0.1944 | | 0.9023 | 0.22 | 100 | 1.2121 | 0.5723 | 0.3083 | 0.3496 | 0.3189 | | 0.5924 | 0.33 | 150 | 0.9662 | 0.6667 | 0.3919 | 0.4265 | 0.4037 | | 0.4218 | 0.44 | 200 | 0.4848 | 0.8813 | 0.6427 | 0.6492 | 0.6413 | | 0.2025 | 0.55 | 250 | 0.3807 | 0.9021 | 0.6677 | 0.6538 | 0.6829 | | 0.1609 | 0.66 | 300 | 0.3360 | 0.9147 | 0.6763 | 0.6727 | 0.6822 | | 0.2035 | 0.76 | 350 | 0.3705 | 0.8991 | 0.6711 | 0.6589 | 0.6838 | | 0.1208 | 0.87 | 400 | 0.2140 | 0.9565 | 0.8218 | 0.8137 | 0.8323 | | 0.1313 | 0.98 | 450 | 0.6818 | 0.8704 | 0.6779 | 0.7179 | 0.6859 | | 0.1576 | 1.09 | 500 | 0.2508 | 0.9212 | 0.7443 | 0.7888 | 0.7311 | | 0.0593 | 1.2 | 550 | 0.2091 | 0.9552 | 0.8193 | 0.8179 | 0.8227 | | 0.0705 | 1.31 | 600 | 0.2010 | 0.9552 | 0.8154 | 0.8091 | 0.8225 | | 0.0637 | 1.42 | 650 | 0.1985 | 0.9573 | 0.8187 | 0.8115 | 0.8275 | | 0.0619 | 1.53 | 700 | 0.2306 | 0.9541 | 0.8241 | 0.8194 | 0.8301 | | 0.0582 | 1.64 | 750 | 0.2001 | 0.9609 | 0.8280 | 0.8250 | 0.8320 | | 0.1132 | 1.75 | 800 | 0.1439 | 0.9680 | 0.8324 | 0.8284 | 0.8367 | | 0.0416 | 1.86 | 850 | 0.1558 | 0.9680 | 0.8333 | 0.8301 | 0.8369 | | 0.0371 | 1.97 | 900 | 0.2242 | 0.9595 | 0.8280 | 0.8235 | 0.8345 | | 0.0428 | 2.07 | 950 | 0.1907 | 0.9617 | 0.8303 | 0.8262 | 0.8356 | | 0.0388 | 2.18 | 1000 | 0.1784 | 0.9658 | 0.8319 | 0.8266 | 0.8383 | | 0.0335 | 2.29 | 1050 | 0.1735 | 0.9675 | 0.8323 | 0.8266 | 0.8390 | | 0.0361 | 2.4 | 1100 | 0.1921 | 0.9636 | 0.8283 | 0.8219 | 0.8360 | | 0.0126 | 2.51 | 1150 | 0.2200 | 0.9614 | 0.8294 | 0.8274 | 0.8327 | | 0.003 | 2.62 | 1200 | 0.2251 | 0.9614 | 0.8296 | 0.8262 | 0.8346 | | 0.0029 | 2.73 | 1250 | 0.1750 | 0.9694 | 0.8348 | 0.8314 | 0.8388 | | 0.0137 | 2.84 | 1300 | 0.1775 | 0.9686 | 0.8345 | 0.8300 | 0.8397 | | 0.0184 | 2.95 | 1350 | 0.1860 | 0.9675 | 0.8337 | 0.8293 | 0.8391 | ### Framework versions - Transformers 4.39.3 - Pytorch 2.1.2 - Datasets 2.18.0 - Tokenizers 0.15.2
kevinkawchak/NousResearch-Hermes-2-Pro-Llama-3-8B-Molecule16
kevinkawchak
2024-05-08T05:55:37Z
8
1
transformers
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "dataset:zjunlp/Mol-Instructions", "base_model:NousResearch/Hermes-2-Pro-Llama-3-8B", "base_model:finetune:NousResearch/Hermes-2-Pro-Llama-3-8B", "license:llama3", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-05-06T07:46:16Z
--- language: - en license: llama3 tags: - text-generation-inference - transformers - unsloth - llama - trl base_model: NousResearch/Hermes-2-Pro-Llama-3-8B datasets: - zjunlp/Mol-Instructions --- - **Developed by:** kevinkawchak - **License:** llama3 - **Finetuned from model :** NousResearch/Hermes-2-Pro-Llama-3-8B - **Finetuned using dataset :** zjunlp/Mol-Instructions, cc-by-4.0 - **Dataset identification:** Molecule-oriented Instructions - **Dataset function:** Description guided molecule design ## May 07, 2024: Additional Fine-tunings, Built with Meta Llama 3 <br> 1) gradientai/Llama-3-8B-Instruct-Gradient-1048k [Model](https://huggingface.co/gradientai/Llama-3-8B-Instruct-Gradient-1048k) <br> Llama 3 8B update: 1040K context length from 8K, and highest RAM consumption<br> "What is the structure for adenine?" Verbose SELFIES structure, but logical<br> [Fine-tuned](https://huggingface.co/kevinkawchak/gradientai-Llama-3-8B-Instruct-Gradient-1048k-Molecule16) on Mol-Instructions, float16, [GitHub](https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/Llama-3-8B-Instruct-Gradient-1048k-Molecule.ipynb), 610 seconds, A100 40GB <br> 2) NousResearch/Hermes-2-Pro-Llama-3-8B [Model](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B)<br> Llama 3 8B update: Cleaned OpenHermes 2.5, new Function Calling, JSON Mode dataset<br> "What is the structure for adenine?" Concise SELFIES structure, but less logical <br> [Fine-tuned](https://huggingface.co/kevinkawchak/NousResearch-Hermes-2-Pro-Llama-3-8B-Molecule16) on Mol-Instructions, float16, [GitHub](https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/Hermes-2-Pro-Llama-3-8B-Molecule.ipynb), 599 seconds, A100 40GB <br> 3) nvidia/Llama3-ChatQA-1.5-8B [Model](https://huggingface.co/nvidia/Llama3-ChatQA-1.5-8B)<br> Llama 3 8B update: ChatQA-1.5 to enhance tabular and arithmetic calculation capability<br> "What is the structure for adenine?" Verbose SELFIES structure and less logical <br> [Fine-tuned](https://huggingface.co/kevinkawchak/nvidia-Llama3-ChatQA-1.5-8B-Molecule16) on Mol-Instructions, float16, [GitHub](https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/Llama3-ChatQA-1.5-8B-Molecule.ipynb), 599 seconds, A100 40GB <br> Responses were verified against the Wikipedia [Adenine](https://en.wikipedia.org/wiki/Adenine) SMILES format and a SMILES to SELFIES python notebook estimated [generator](https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/SMILES%20to%20SELFIES%20estimator.ipynb). <br> Fine-tunings were performed using the Apache-2.0 unsloth 'Alpaca + Llama-3 8b full example' Colab [notebook](https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing). ## Primary Study The following are modifications or improvements to original notebooks. Please refer to the authors' models for the published primary work. [Cover Image](https://drive.google.com/file/d/1J-spZMzLlPxkqfMrPxvtMZiD2_hfcGyr/view?usp=sharing). [META LLAMA 3 COMMUNITY LICENSE AGREEMENT](https://llama.meta.com/llama3/license/). Built with Meta Llama 3. <br> A 4-bit quantization of Meta-Llama-3-8B-Instruct was used to reduce training memory requirements when fine-tuning on the zjunlp/Mol-Instructions dataset. (1-2) In addition, the minimum LoRA rank value was utilized to reduce the overall size of created models. In specific, the molecule-oriented instructions description guided molecule design was implemented to answer general questions and general biochemistry questions. General questions were answered with high accuracy, while biochemistry related questions returned 'SELFIES' structures but with limited accuracy. The notebook featured Torch and Hugging Face libraries using the Unsloth llama-3-8b-Instruct-bnb-4bit quantization model. Training loss decreased steadily from 1.97 to 0.73 over 60 steps. Additional testing regarding the appropriate level of compression or hyperparameter adjustments for accurate SELFIES chemical structures outputs is relevant, as shown in the GitHub notebook for research purposes (3). A 16-bit and reduced 4-bit size were uploaded to Hugging Face. (4-5) Update 04/24: The number of training steps were increased to further decrease loss, while maintaining reduced memory requirements through quantization and reduced size through LoRA. This allowed for significantly improved responses to biochemistry related questions, and were saved at the following LLM Model sizes: [8.03B](https://huggingface.co/kevinkawchak/Meta-Llama-3-8B-Instruct-Molecule16), [4.65B](https://huggingface.co/kevinkawchak/Meta-Llama-3-8B-Instruct-Molecule04). [github](https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/Meta-Llama-3-8B-Instruct-Molecule.ipynb). References: 1) unsloth: https://huggingface.co/unsloth/llama-3-8b-Instruct-bnb-4bit 2) zjunlp: https://huggingface.co/datasets/zjunlp/Mol-Instructions 3) github: https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/Meta-Llama-3-8B-Instruct-Mol.ipynb 4) hugging face: https://huggingface.co/kevinkawchak/Meta-Llama-3-8B-Instruct-LoRA-Mol16 5) hugging face: https://huggingface.co/kevinkawchak/Meta-Llama-3-8B-Instruct-LoRA-Mol04 @inproceedings{fang2023mol, <br> author = {Yin Fang and<br> Xiaozhuan Liang and<br> Ningyu Zhang and<br> Kangwei Liu and<br> Rui Huang and<br> Zhuo Chen and<br> Xiaohui Fan and<br> Huajun Chen},<br> title = {Mol-Instructions: {A} Large-Scale Biomolecular Instruction Dataset<br> for Large Language Models},<br> booktitle = {{ICLR}},<br> publisher = {OpenReview.net},<br> year = {2024},<br> url = {https://openreview.net/pdf?id=Tlsdsb6l9n}}<br> This llama model was trained with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
kevinkawchak/gradientai-Llama-3-8B-Instruct-Gradient-1048k-Molecule16
kevinkawchak
2024-05-08T05:55:15Z
9
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "dataset:zjunlp/Mol-Instructions", "base_model:gradientai/Llama-3-8B-Instruct-Gradient-1048k", "base_model:finetune:gradientai/Llama-3-8B-Instruct-Gradient-1048k", "license:llama3", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2024-05-06T05:42:11Z
--- language: - en license: llama3 tags: - text-generation-inference - transformers - unsloth - llama - trl base_model: gradientai/Llama-3-8B-Instruct-Gradient-1048k datasets: - zjunlp/Mol-Instructions --- - **Developed by:** kevinkawchak - **License:** llama3 - **Finetuned from model :** gradientai/Llama-3-8B-Instruct-Gradient-1048k - **Finetuned using dataset :** zjunlp/Mol-Instructions, cc-by-4.0 - **Dataset identification:** Molecule-oriented Instructions - **Dataset function:** Description guided molecule design ## May 07, 2024: Additional Fine-tunings, Built with Meta Llama 3 <br> 1) gradientai/Llama-3-8B-Instruct-Gradient-1048k [Model](https://huggingface.co/gradientai/Llama-3-8B-Instruct-Gradient-1048k) <br> Llama 3 8B update: 1040K context length from 8K, and highest RAM consumption<br> "What is the structure for adenine?" Verbose SELFIES structure, but logical<br> [Fine-tuned](https://huggingface.co/kevinkawchak/gradientai-Llama-3-8B-Instruct-Gradient-1048k-Molecule16) on Mol-Instructions, float16, [GitHub](https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/Llama-3-8B-Instruct-Gradient-1048k-Molecule.ipynb), 610 seconds, A100 40GB <br> 2) NousResearch/Hermes-2-Pro-Llama-3-8B [Model](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B)<br> Llama 3 8B update: Cleaned OpenHermes 2.5, new Function Calling, JSON Mode dataset<br> "What is the structure for adenine?" Concise SELFIES structure, but less logical <br> [Fine-tuned](https://huggingface.co/kevinkawchak/NousResearch-Hermes-2-Pro-Llama-3-8B-Molecule16) on Mol-Instructions, float16, [GitHub](https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/Hermes-2-Pro-Llama-3-8B-Molecule.ipynb), 599 seconds, A100 40GB <br> 3) nvidia/Llama3-ChatQA-1.5-8B [Model](https://huggingface.co/nvidia/Llama3-ChatQA-1.5-8B)<br> Llama 3 8B update: ChatQA-1.5 to enhance tabular and arithmetic calculation capability<br> "What is the structure for adenine?" Verbose SELFIES structure and less logical <br> [Fine-tuned](https://huggingface.co/kevinkawchak/nvidia-Llama3-ChatQA-1.5-8B-Molecule16) on Mol-Instructions, float16, [GitHub](https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/Llama3-ChatQA-1.5-8B-Molecule.ipynb), 599 seconds, A100 40GB <br> Responses were verified against the Wikipedia [Adenine](https://en.wikipedia.org/wiki/Adenine) SMILES format and a SMILES to SELFIES python notebook estimated [generator](https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/SMILES%20to%20SELFIES%20estimator.ipynb). <br> Fine-tunings were performed using the Apache-2.0 unsloth 'Alpaca + Llama-3 8b full example' Colab [notebook](https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing). ## Primary Study The following are modifications or improvements to original notebooks. Please refer to the authors' models for the published primary work. [Cover Image](https://drive.google.com/file/d/1J-spZMzLlPxkqfMrPxvtMZiD2_hfcGyr/view?usp=sharing). [META LLAMA 3 COMMUNITY LICENSE AGREEMENT](https://llama.meta.com/llama3/license/). Built with Meta Llama 3. <br> A 4-bit quantization of Meta-Llama-3-8B-Instruct was used to reduce training memory requirements when fine-tuning on the zjunlp/Mol-Instructions dataset. (1-2) In addition, the minimum LoRA rank value was utilized to reduce the overall size of created models. In specific, the molecule-oriented instructions description guided molecule design was implemented to answer general questions and general biochemistry questions. General questions were answered with high accuracy, while biochemistry related questions returned 'SELFIES' structures but with limited accuracy. The notebook featured Torch and Hugging Face libraries using the Unsloth llama-3-8b-Instruct-bnb-4bit quantization model. Training loss decreased steadily from 1.97 to 0.73 over 60 steps. Additional testing regarding the appropriate level of compression or hyperparameter adjustments for accurate SELFIES chemical structures outputs is relevant, as shown in the GitHub notebook for research purposes (3). A 16-bit and reduced 4-bit size were uploaded to Hugging Face. (4-5) Update 04/24: The number of training steps were increased to further decrease loss, while maintaining reduced memory requirements through quantization and reduced size through LoRA. This allowed for significantly improved responses to biochemistry related questions, and were saved at the following LLM Model sizes: [8.03B](https://huggingface.co/kevinkawchak/Meta-Llama-3-8B-Instruct-Molecule16), [4.65B](https://huggingface.co/kevinkawchak/Meta-Llama-3-8B-Instruct-Molecule04). [github](https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/Meta-Llama-3-8B-Instruct-Molecule.ipynb). References: 1) unsloth: https://huggingface.co/unsloth/llama-3-8b-Instruct-bnb-4bit 2) zjunlp: https://huggingface.co/datasets/zjunlp/Mol-Instructions 3) github: https://github.com/kevinkawchak/Medical-Quantum-Machine-Learning/blob/main/Code/Drug%20Discovery/Meta-Llama-3/Meta-Llama-3-8B-Instruct-Mol.ipynb 4) hugging face: https://huggingface.co/kevinkawchak/Meta-Llama-3-8B-Instruct-LoRA-Mol16 5) hugging face: https://huggingface.co/kevinkawchak/Meta-Llama-3-8B-Instruct-LoRA-Mol04 @inproceedings{fang2023mol, <br> author = {Yin Fang and<br> Xiaozhuan Liang and<br> Ningyu Zhang and<br> Kangwei Liu and<br> Rui Huang and<br> Zhuo Chen and<br> Xiaohui Fan and<br> Huajun Chen},<br> title = {Mol-Instructions: {A} Large-Scale Biomolecular Instruction Dataset<br> for Large Language Models},<br> booktitle = {{ICLR}},<br> publisher = {OpenReview.net},<br> year = {2024},<br> url = {https://openreview.net/pdf?id=Tlsdsb6l9n}}<br> This llama model was trained with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
T3Q-LLM/T3Q-LLM1-CV-v2.0
T3Q-LLM
2024-05-08T05:55:12Z
2,250
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2024-05-02T12:23:02Z
--- library_name: transformers license: apache-2.0 --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation hf (pretrained=T3Q-LLM/T3Q-LLM1-v1.0), limit: None, provide_description: False, num_fewshot: 0, batch_size: None | Task |Version| Metric |Value | |Stderr| |----------------|------:|--------|-----:|---|-----:| |kobest_boolq | 0|acc |0.9437|± |0.0062| | | |macro_f1|0.9437|± |0.0062| |kobest_copa | 0|acc |0.7720|± |0.0133| | | |macro_f1|0.7718|± |0.0133| |kobest_hellaswag| 0|acc |0.5060|± |0.0224| | | |acc_norm|0.5320|± |0.0223| | | |macro_f1|0.5035|± |0.0224| |kobest_sentineg | 0|acc |0.8287|± |0.0189| | | |macro_f1|0.8275|± |0.0191|