ryan0712 commited on
Commit
8823e03
1 Parent(s): 748ff94

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -12
README.md CHANGED
@@ -13,26 +13,61 @@ This is an unofficial implementation of "[AlpaGasus: Training a better Alpaca wi
13
  - **License**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
14
 
15
 
16
-
17
- ### Benchmark Metrics
18
- | Metric | Value |
19
- |-----------------------|-------|
20
- | MMLU | 55.27 |
21
- | ARC | 61.09 |
22
- | HellaSwag | 82.46 |
23
- | TruthfulQA | 38.53 |
24
- | Avg. | 59.34 |
25
-
26
  ### Training Dataset
27
 
28
  "StudentLLM/Alpagasus-2-13b-QLoRA-merged" used [gpt4life](https://github.com/gpt4life/alpagasus)'s gpt-3.5-turbo filtered dataset, 'alpaca_t45.json'.
29
  Configuration of the dataset is as follows:
30
 
31
- ### Prompt Template
32
  ```
33
- ### Instruction:
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
 
35
  <prompt> (without the <>)
36
 
 
 
 
37
  ### Response:
38
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  - **License**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
14
 
15
 
 
 
 
 
 
 
 
 
 
 
16
  ### Training Dataset
17
 
18
  "StudentLLM/Alpagasus-2-13b-QLoRA-merged" used [gpt4life](https://github.com/gpt4life/alpagasus)'s gpt-3.5-turbo filtered dataset, 'alpaca_t45.json'.
19
  Configuration of the dataset is as follows:
20
 
 
21
  ```
22
+ {
23
+ 'instruction': Give the instruction describing the question.
24
+ 'input': Occasionally present, detailed instructions accompany the question if available.
25
+ 'output': Give answers to questions.
26
+ }
27
+ .
28
+ .
29
+ .
30
+
31
+ ```
32
+
33
+ ### Prompt Template: Alpaca style prompt
34
+ ```
35
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
36
 
37
+ ### Instruction:
38
  <prompt> (without the <>)
39
 
40
+ ### Input:
41
+ <prompt> (if input exists)
42
+
43
  ### Response:
44
  ```
45
+
46
+ ### Fine-tuning Procedure
47
+ Our model was finetuned using QLoRA on single A100 80GB GPU. Training details are described in [repo](https://github.com/gauss5930/AlpaGasus2-QLoRA).
48
+
49
+ ### Benchmark Metrics
50
+ "StudentLLM/Alpagasus-2-13b-QLoRA-merged" model performance is uploaded on Huggingface's [OpenLLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). Model was evaluated on the tasks specified in HF's Open LLM Leaderboard(ARC, HellaSwag, MMLU, TruthfulQA).
51
+
52
+ | Metric | Value |
53
+ |-----------------------|-------|
54
+ | Avg. | 59.34 |
55
+ | MMLU | 55.27 |
56
+ | ARC | 61.09 |
57
+ | HellaSwag | 82.46 |
58
+ | TruthfulQA | 38.53 |
59
+
60
+ ### LLM Evaluation
61
+
62
+ ### Fine-tuning Procedure
63
+ Our mod
64
+
65
+ ### Citations
66
+ ```bibtex
67
+ @article{chen2023alpagasus,
68
+ title={AlpaGasus: Training a Better Alpaca with Fewer Data},
69
+ author={Lichang Chen, Shiyang Li, Jun Yan, Hai Wang, Kalpa Gunaratna, Vikas Yadav, Zheng Tang, Vijay Srinivasan, Tianyi Zhou, Heng Huang, Hongxia Jin},
70
+ journal={arXiv preprint arXiv:2307.08701},
71
+ year={2023}
72
+ }
73
+ ```