Phi-2-QLora / README.md
JamieAi33's picture
Update README.md
3c85c02 verified
---
base_model: microsoft/phi-2
library_name: peft
license: apache-2.0
datasets:
- neil-code/dialogsum-test
language:
- en
metrics:
- bleu
pipeline_tag: question-answering
tags:
- QuestionAnswering
- legal
- finan
- chem
- biology
---
license: apache-2.0
language:
- en
metrics:
- rouge
base_model:
- microsoft/phi-2
pipeline_tag: question-answering
---
This repo containes the last checkpoint of my fine tuned model. Click this link to go the final model
https://huggingface.co/JamieAi33/Phi-2_PEFT
# Model Card for PEFT-Fine-Tuned Model
This model card documents a PEFT-fine-tuned version of `microsoft/phi-2` for question-answering tasks. The PEFT fine-tuning improved the model's performance, as detailed in the evaluation section.
## Model Details
### Model Description
- **Developed by:** JamieAi33
- **Finetuned from model:** `microsoft/phi-2`
- **Model type:** PEFT fine-tuned transformer
- **Language(s) (NLP):** English
- **License:** Apache 2.0
The base model `microsoft/phi-2` was adapted using Parameter-Efficient Fine-Tuning (PEFT) for question-answering tasks. The training process focused on improving performance metrics while keeping computational costs low.
---
### Model Sources
- **Repository:** https://huggingface.co/JamieAi33/Phi-2-QLora
- **Paper:** [Optional: Add a reference to PEFT or any relevant paper]
- **Demo:** [Optional: Link to your Hugging Face Space or demo]
---
## Uses
### Direct Use
This model can be used out-of-the-box for question-answering tasks.
### Downstream Use
The model can be fine-tuned further on domain-specific datasets for improved performance.
### Out-of-Scope Use
Avoid using this model for tasks outside question-answering or where fairness, bias, and ethical considerations are critical without further validation.
---
## Bias, Risks, and Limitations
Users should be aware that:
- The model is trained on publicly available data and may inherit biases present in the training data.
- It is optimized for English and may perform poorly in other languages.
---
## How to Get Started with the Model
Here鈥檚 an example of loading the model:
```python
from transformers import AutoModel
from peft import PeftModel
base_model = AutoModel.from_pretrained("microsoft/phi-2")
adapter_model = PeftModel.from_pretrained(base_model, "JamieAi33/Phi-2-QLora")
# Model Name: PEFT Fine-Tuned `microsoft/phi-2`
This repository contains a PEFT fine-tuned version of the `microsoft/phi-2` model for question-answering tasks. The fine-tuning process leveraged Parameter-Efficient Fine-Tuning (PEFT) techniques to achieve improved performance.
---
## Metrics
The model's performance was evaluated using the ROUGE metric. Below are the results:
| **Metric** | **Original Model** | **PEFT Model** | **Absolute Improvement** |
|-----------------|--------------------|----------------|---------------------------|
| **ROUGE-1** | 29.76% | 44.51% | +14.75% |
| **ROUGE-2** | 10.76% | 15.68% | +4.92% |
| **ROUGE-L** | 21.69% | 30.95% | +9.25% |
| **ROUGE-Lsum** | 22.75% | 31.49% | +8.74% |
---
## Training Configuration
| Hyperparameter | Value |
|-----------------------|-------------------------|
| **Batch Size** | 1 |
| **Learning Rate** | 2e-4 |
| **Max Steps** | 1000 |
| **Optimizer** | Paged AdamW (8-bit) |
| **Logging Steps** | 25 |
| **Evaluation Steps** | 25 |
| **Gradient Checkpointing** | Enabled |