akumaburn commited on
Commit
e6ac807
·
verified ·
1 Parent(s): 971f8b5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -2
README.md CHANGED
@@ -7,15 +7,67 @@ tags:
7
  - transformers
8
  - unsloth
9
  - llama
10
- - gguf
 
11
  base_model: unsloth/llama-3-8b-bnb-4bit
 
 
12
  ---
13
 
14
- # Uploaded model
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
  - **Developed by:** akumaburn
17
  - **License:** apache-2.0
18
  - **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
21
 
 
7
  - transformers
8
  - unsloth
9
  - llama
10
+ - trl
11
+ - sft
12
  base_model: unsloth/llama-3-8b-bnb-4bit
13
+ datasets:
14
+ - yahma/alpaca-cleaned
15
  ---
16
 
17
+ # Alpaca-Llama-3-8B
18
+
19
+
20
+ - **Fine Tuned using dataset:** https://huggingface.co/datasets/yahma/alpaca-cleaned
21
+ - **Epoch Count:** 1
22
+ - **Step Count:** 6,470/6,470
23
+ - **Batch Size:** 2
24
+ - **Gradient Accumulation Steps:** 4
25
+ - **Context Size:** 8192
26
+ - **Num examples:** 51,760
27
+ - **Trainable Parameters:** 41,943,040
28
+ - **Learning Rate:** 0.00001
29
+ - **Training Loss:** 0.960000
30
+ - **Fined Tuned using:** Google Colab Pro (Nvidia T4 runtime)
31
 
32
  - **Developed by:** akumaburn
33
  - **License:** apache-2.0
34
  - **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
35
+ - **Prompt Format:** Alpaca System (https://libertai.io/apis/text-generation/prompting.html)
36
+
37
+ Full model can be found in https://huggingface.co/akumaburn/Alpaca-Llama-3-8B
38
+
39
+ mistral-7b-openorca.Q8_0.gguf:
40
+ - **MMLU-Test:** Final result: **41.5836 +/- 0.4174**
41
+ - **Arc-Easy:** Final result: 72.6316 +/- 1.8691
42
+ - **Truthful QA:** Final result: **32.0685 +/- 1.6339**
43
+ - **Arc-Challenge:** Final result: 48.8294 +/- 2.8956
44
+
45
+ llama-3-8b-bnb-4bit.Q8_0.gguf:
46
+ - **MMLU-Test:** Final result: 40.4074 +/- 0.4156
47
+ - **Arc-Easy:** Final result: 73.8596 +/- 1.8421
48
+ - **Truthful QA:** Final result: 26.6830 +/- 1.5484
49
+ - **Arc-Challenge:** Final result: 46.8227 +/- 2.8906
50
+
51
+ Open_Orca_Llama-3-8B-unsloth.Q8_0.gguf:
52
+ - **MMLU-Test:** Final result: 39.3818 +/- 0.4138
53
+ - **Arc-Easy:** Final result: 67.3684 +/- 1.9656
54
+ - **Truthful QA:** Final result: 29.0086 +/- 1.5886
55
+ - **Arc-Challenge:** Final result: 42.1405 +/- 2.8604
56
+
57
+ **Alpaca-Llama-3-8B-GGUF-unsloth.Q8_0.gguf**:
58
+ - **MMLU-Test:** Final result: 40.6441 +/- 0.4160
59
+ - **Arc-Easy:** Final result: **77.5439 +/- 1.7494**
60
+ - **Truthful QA:** Final result: 29.7430 +/- 1.6003
61
+ - **Arc-Challenge:** Final result: **50.5017 +/- 2.8963**
62
+
63
+ Meta-Llama-3-8B.Q8_0.gguf:
64
+ - **MMLU-Test:** Final result: 40.8664 +/- 0.4163
65
+ - **Arc-Easy:** Final result: 74.3860 +/- 1.8299
66
+ - **Truthful QA:** Final result: 28.6414 +/- 1.5826
67
+ - **Arc-Challenge:** Final result: 47.1572 +/- 2.8917
68
+
69
+ Llama.cpp Options For Testing:
70
+ --samplers "tfs;typical;temp" --draft 32 --ctx-size 8192 --temp 0.82 --tfs 0.8 --typical 1.1 --repeat-last-n 512 --batch-size 8192 --repeat-penalty 1.0 --n-gpu-layers 100 --threads 12
71
 
72
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
73