| | --- |
| | base_model: meta-llama/Meta-Llama-3-8B |
| | library_name: peft |
| | license: llama3 |
| | datasets: |
| | - lewtun/github-issues |
| | language: |
| | - en |
| | pipeline_tag: summarization |
| | --- |
| | |
| | # Model Card: LoRA-LLaMA3-8B-GitHub-Summarizer |
| |
|
| | This repository provides LoRA adapter weights fine-tuned on top of Meta’s LLaMA-3-8B model for the task of summarizing GitHub issues and discussions. The model was trained on a curated dataset of open-source GitHub issues to produce concise, readable, and technically accurate summaries. |
| |
|
| | ## Model Details |
| |
|
| | ### Model Description |
| |
|
| | - **Developed by:** Saramsh Gautam (Louisiana State University) |
| | - **Model type:** LoRA adapter weights |
| | - **Language(s):** English |
| | - **License:** llama (must comply with Meta's license) |
| | - **Fine-tuned from model:** `meta-llama/Meta-Llama-3-8B` |
| | - **Library used:** PEFT (LoRA) with Hugging Face Transformers |
| |
|
| | ### Model Sources |
| |
|
| | - **Base model:** [Meta-LLaMA-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) |
| | - **Repository:** [link to this repo]() |
| |
|
| | ## Uses |
| |
|
| | ### Direct Use |
| |
|
| | These adapter weights must be merged with the base LLaMA-3-8B model using PEFT or Hugging Face’s `PeftModel` wrapper. |
| |
|
| | Example use case: |
| |
|
| | ```python |
| | from transformers import AutoTokenizer, AutoModelForCausalLM |
| | from peft import PeftModel, PeftConfig |
| | |
| | base_model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B") |
| | model = PeftModel.from_pretrained(base_model, "saramshgautam/lora-llama-8b-github") |
| | tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B") |
| | ``` |
| |
|
| | ### Intended USe |
| |
|
| | - Research in summarization of technical conversations |
| |
|
| | - Augmenting code review and issue tracking pipelines |
| |
|
| | - Studying model adaptation via parameter-efficient fine-tuning |
| |
|
| | ### Out-of-Scope Use |
| |
|
| | - Commercial applications (restricted by Meta’s LLaMA license) |
| |
|
| | - General-purpose conversation or chatbot use (model optimized for summarization) |
| |
|
| | ## Bias, Risks, and Limitations |
| |
|
| | - The model inherits biases from both the base LLaMA-3 model and the GitHub dataset. It may underperform on non-technical content or multilingual issues. |
| |
|
| | ## Recommendations |
| |
|
| | Use only for academic or non-commercial research. Evaluate responsibly before using in production or public-facing tools. |
| |
|
| | ## How to Get Started with the Model |
| |
|
| | See the example in “Direct Use” above. You must separately download the base model from Meta and load the LoRA adapters from this repo. |
| |
|
| | ## Training Details |
| |
|
| | ### Training Data |
| |
|
| | - Source: Hugging Face lewtun/github-issues |
| | - Description: Contains 3,000+ GitHub issues and comments from popular open-source repositories. |
| |
|
| | ## Training Procedure |
| |
|
| | - LoRA with PEFT |
| | - 4-bit quantized training using bitsandbytes |
| | - Mixed precision: bf16 |
| | - Batch size: 8 |
| | - Epochs: 3 |
| | - Optimizer: AdamW |
| |
|
| | ### Evaluation |
| |
|
| | ## Metrics |
| |
|
| | ROUGE-1, ROUGE-2, ROUGE-L, ROUGE-Lsum on a 500-issue test set |
| |
|
| | ## Results |
| |
|
| | | Metric | Score | |
| | | ---------- | ----- | |
| | | ROUGE-1 | 0.706 | |
| | | ROUGE-2 | 0.490 | |
| | | ROUGE-L | 0.570 | |
| | | ROUGE-Lsum | 0.582 | |
| |
|
| | --- |
| |
|
| | ## Environmental Impact |
| |
|
| | - **Hardware Type:** 4×A100 GPUs (university HPC cluster) |
| | - **Training Hours:** ~4 hours |
| | - **Carbon Estimate:** ~10.2 kg CO₂eq |
| | _(estimated via [ML CO2 calculator](https://mlco2.github.io/impact))_ |
| |
|
| | --- |
| |
|
| | ## Citation |
| |
|
| | **APA:** |
| |
|
| | Gautam, S. (2025). _LoRA-LLaMA3-8B-GitHub-Summarizer: Adapter weights for summarizing GitHub issues using LLaMA 3_. Hugging Face. https://huggingface.co/saramshgautam/lora-llama-8b-github |
| |
|
| | **BibTeX:** |
| |
|
| | ```bibtex |
| | @misc{gautam2025lora, |
| | title={LoRA-LLaMA3-8B-GitHub-Summarizer}, |
| | author={Gautam, Saramsh}, |
| | year={2025}, |
| | howpublished={\url{https://huggingface.co/saramshgautam/lora-llama-8b-github}}, |
| | note={Fine-tuned adapter weights using LoRA on Meta-LLaMA-3-8B} |
| | } |
| | ``` |
| |
|
| | --- |
| |
|
| | ## Contact |
| |
|
| | - **Author:** Saramsh Gautam |
| | - **Affiliation:** Louisiana State University |
| | - **Email:** [your email] |
| | - **Hugging Face profile:** [https://huggingface.co/saramshgautam](https://huggingface.co/saramshgautam) |
| |
|
| | --- |
| |
|
| | ## Framework Versions |
| |
|
| | - **PEFT:** 0.15.2 |
| | - **Transformers:** 4.40.0 |
| | - **Bitsandbytes:** 0.41.3 |
| | - **Datasets:** 2.18.0 |