|
--- |
|
base_model: meta-llama/Meta-Llama-3-8B |
|
library_name: peft |
|
license: llama3 |
|
datasets: |
|
- lewtun/github-issues |
|
language: |
|
- en |
|
pipeline_tag: summarization |
|
--- |
|
|
|
# Model Card: LoRA-LLaMA3-8B-GitHub-Summarizer |
|
|
|
This repository provides LoRA adapter weights fine-tuned on top of Meta’s LLaMA-3-8B model for the task of summarizing GitHub issues and discussions. The model was trained on a curated dataset of open-source GitHub issues to produce concise, readable, and technically accurate summaries. |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
- **Developed by:** Saramsh Gautam (Louisiana State University) |
|
- **Model type:** LoRA adapter weights |
|
- **Language(s):** English |
|
- **License:** llama (must comply with Meta's license) |
|
- **Fine-tuned from model:** `meta-llama/Meta-Llama-3-8B` |
|
- **Library used:** PEFT (LoRA) with Hugging Face Transformers |
|
|
|
### Model Sources |
|
|
|
- **Base model:** [Meta-LLaMA-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) |
|
- **Repository:** [link to this repo]() |
|
|
|
## Uses |
|
|
|
### Direct Use |
|
|
|
These adapter weights must be merged with the base LLaMA-3-8B model using PEFT or Hugging Face’s `PeftModel` wrapper. |
|
|
|
Example use case: |
|
|
|
```python |
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
from peft import PeftModel, PeftConfig |
|
|
|
base_model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B") |
|
model = PeftModel.from_pretrained(base_model, "saramshgautam/lora-llama-8b-github") |
|
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B") |
|
``` |
|
|
|
### Intended USe |
|
|
|
- Research in summarization of technical conversations |
|
|
|
- Augmenting code review and issue tracking pipelines |
|
|
|
- Studying model adaptation via parameter-efficient fine-tuning |
|
|
|
### Out-of-Scope Use |
|
|
|
- Commercial applications (restricted by Meta’s LLaMA license) |
|
|
|
- General-purpose conversation or chatbot use (model optimized for summarization) |
|
|
|
## Bias, Risks, and Limitations |
|
|
|
- The model inherits biases from both the base LLaMA-3 model and the GitHub dataset. It may underperform on non-technical content or multilingual issues. |
|
|
|
## Recommendations |
|
|
|
Use only for academic or non-commercial research. Evaluate responsibly before using in production or public-facing tools. |
|
|
|
## How to Get Started with the Model |
|
|
|
See the example in “Direct Use” above. You must separately download the base model from Meta and load the LoRA adapters from this repo. |
|
|
|
## Training Details |
|
|
|
### Training Data |
|
|
|
- Source: Hugging Face lewtun/github-issues |
|
- Description: Contains 3,000+ GitHub issues and comments from popular open-source repositories. |
|
|
|
## Training Procedure |
|
|
|
- LoRA with PEFT |
|
- 4-bit quantized training using bitsandbytes |
|
- Mixed precision: bf16 |
|
- Batch size: 8 |
|
- Epochs: 3 |
|
- Optimizer: AdamW |
|
|
|
### Evaluation |
|
|
|
## Metrics |
|
|
|
ROUGE-1, ROUGE-2, ROUGE-L, ROUGE-Lsum on a 500-issue test set |
|
|
|
## Results |
|
|
|
| Metric | Score | |
|
| ---------- | ----- | |
|
| ROUGE-1 | 0.706 | |
|
| ROUGE-2 | 0.490 | |
|
| ROUGE-L | 0.570 | |
|
| ROUGE-Lsum | 0.582 | |
|
|
|
--- |
|
|
|
## Environmental Impact |
|
|
|
- **Hardware Type:** 4×A100 GPUs (university HPC cluster) |
|
- **Training Hours:** ~4 hours |
|
- **Carbon Estimate:** ~10.2 kg CO₂eq |
|
_(estimated via [ML CO2 calculator](https://mlco2.github.io/impact))_ |
|
|
|
--- |
|
|
|
## Citation |
|
|
|
**APA:** |
|
|
|
Gautam, S. (2025). _LoRA-LLaMA3-8B-GitHub-Summarizer: Adapter weights for summarizing GitHub issues using LLaMA 3_. Hugging Face. https://huggingface.co/saramshgautam/lora-llama-8b-github |
|
|
|
**BibTeX:** |
|
|
|
```bibtex |
|
@misc{gautam2025lora, |
|
title={LoRA-LLaMA3-8B-GitHub-Summarizer}, |
|
author={Gautam, Saramsh}, |
|
year={2025}, |
|
howpublished={\url{https://huggingface.co/saramshgautam/lora-llama-8b-github}}, |
|
note={Fine-tuned adapter weights using LoRA on Meta-LLaMA-3-8B} |
|
} |
|
``` |
|
|
|
--- |
|
|
|
## Contact |
|
|
|
- **Author:** Saramsh Gautam |
|
- **Affiliation:** Louisiana State University |
|
- **Email:** [your email] |
|
- **Hugging Face profile:** [https://huggingface.co/saramshgautam](https://huggingface.co/saramshgautam) |
|
|
|
--- |
|
|
|
## Framework Versions |
|
|
|
- **PEFT:** 0.15.2 |
|
- **Transformers:** 4.40.0 |
|
- **Bitsandbytes:** 0.41.3 |
|
- **Datasets:** 2.18.0 |