metadata
license: other
language:
- en
Model Details
This is an unofficial implementation of "AlpaGasus: Training a better Alpaca with Fewer Data." with LLaMA2 & QLoRA! Training code is available at our repo.
- Developed by: Yunsang Yoo and Hyunwoo Ko
- Model type: Auto-regressive model
- Language(s): English
- Base Model: meta-llama/Llama-2-13b-hf
- License: Non-Commercial Creative Commons license (CC BY-NC-4.0)
Benchmark Metrics
Metric | Value |
---|---|
MMLU | 55.27 |
ARC | 61.09 |
HellaSwag | 82.46 |
TruthfulQA | 38.53 |
Avg. | 59.34 |
Training Dataset
"StudentLLM/Alpagasus-2-13b-QLoRA-merged" used gpt4life's gpt-3.5-turbo filtered dataset, 'alpaca_t45.json'. Configuration of the dataset is as follows:
Prompt Template
### Instruction:
<prompt> (without the <>)
### Response: