Update README.md
Browse files
README.md
CHANGED
@@ -13,26 +13,61 @@ This is an unofficial implementation of "[AlpaGasus: Training a better Alpaca wi
|
|
13 |
- **License**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
|
14 |
|
15 |
|
16 |
-
|
17 |
-
### Benchmark Metrics
|
18 |
-
| Metric | Value |
|
19 |
-
|-----------------------|-------|
|
20 |
-
| MMLU | 55.27 |
|
21 |
-
| ARC | 61.09 |
|
22 |
-
| HellaSwag | 82.46 |
|
23 |
-
| TruthfulQA | 38.53 |
|
24 |
-
| Avg. | 59.34 |
|
25 |
-
|
26 |
### Training Dataset
|
27 |
|
28 |
"StudentLLM/Alpagasus-2-13b-QLoRA-merged" used [gpt4life](https://github.com/gpt4life/alpagasus)'s gpt-3.5-turbo filtered dataset, 'alpaca_t45.json'.
|
29 |
Configuration of the dataset is as follows:
|
30 |
|
31 |
-
### Prompt Template
|
32 |
```
|
33 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
|
|
|
35 |
<prompt> (without the <>)
|
36 |
|
|
|
|
|
|
|
37 |
### Response:
|
38 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
- **License**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
|
14 |
|
15 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
### Training Dataset
|
17 |
|
18 |
"StudentLLM/Alpagasus-2-13b-QLoRA-merged" used [gpt4life](https://github.com/gpt4life/alpagasus)'s gpt-3.5-turbo filtered dataset, 'alpaca_t45.json'.
|
19 |
Configuration of the dataset is as follows:
|
20 |
|
|
|
21 |
```
|
22 |
+
{
|
23 |
+
'instruction': Give the instruction describing the question.
|
24 |
+
'input': Occasionally present, detailed instructions accompany the question if available.
|
25 |
+
'output': Give answers to questions.
|
26 |
+
}
|
27 |
+
.
|
28 |
+
.
|
29 |
+
.
|
30 |
+
|
31 |
+
```
|
32 |
+
|
33 |
+
### Prompt Template: Alpaca style prompt
|
34 |
+
```
|
35 |
+
Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
36 |
|
37 |
+
### Instruction:
|
38 |
<prompt> (without the <>)
|
39 |
|
40 |
+
### Input:
|
41 |
+
<prompt> (if input exists)
|
42 |
+
|
43 |
### Response:
|
44 |
```
|
45 |
+
|
46 |
+
### Fine-tuning Procedure
|
47 |
+
Our model was finetuned using QLoRA on single A100 80GB GPU. Training details are described in [repo](https://github.com/gauss5930/AlpaGasus2-QLoRA).
|
48 |
+
|
49 |
+
### Benchmark Metrics
|
50 |
+
"StudentLLM/Alpagasus-2-13b-QLoRA-merged" model performance is uploaded on Huggingface's [OpenLLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). Model was evaluated on the tasks specified in HF's Open LLM Leaderboard(ARC, HellaSwag, MMLU, TruthfulQA).
|
51 |
+
|
52 |
+
| Metric | Value |
|
53 |
+
|-----------------------|-------|
|
54 |
+
| Avg. | 59.34 |
|
55 |
+
| MMLU | 55.27 |
|
56 |
+
| ARC | 61.09 |
|
57 |
+
| HellaSwag | 82.46 |
|
58 |
+
| TruthfulQA | 38.53 |
|
59 |
+
|
60 |
+
### LLM Evaluation
|
61 |
+
|
62 |
+
### Fine-tuning Procedure
|
63 |
+
Our mod
|
64 |
+
|
65 |
+
### Citations
|
66 |
+
```bibtex
|
67 |
+
@article{chen2023alpagasus,
|
68 |
+
title={AlpaGasus: Training a Better Alpaca with Fewer Data},
|
69 |
+
author={Lichang Chen, Shiyang Li, Jun Yan, Hai Wang, Kalpa Gunaratna, Vikas Yadav, Zheng Tang, Vijay Srinivasan, Tianyi Zhou, Heng Huang, Hongxia Jin},
|
70 |
+
journal={arXiv preprint arXiv:2307.08701},
|
71 |
+
year={2023}
|
72 |
+
}
|
73 |
+
```
|