Update README.md
Browse files
README.md
CHANGED
@@ -7,15 +7,67 @@ tags:
|
|
7 |
- transformers
|
8 |
- unsloth
|
9 |
- llama
|
10 |
-
-
|
|
|
11 |
base_model: unsloth/llama-3-8b-bnb-4bit
|
|
|
|
|
12 |
---
|
13 |
|
14 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
- **Developed by:** akumaburn
|
17 |
- **License:** apache-2.0
|
18 |
- **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
19 |
|
20 |
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
|
21 |
|
|
|
7 |
- transformers
|
8 |
- unsloth
|
9 |
- llama
|
10 |
+
- trl
|
11 |
+
- sft
|
12 |
base_model: unsloth/llama-3-8b-bnb-4bit
|
13 |
+
datasets:
|
14 |
+
- yahma/alpaca-cleaned
|
15 |
---
|
16 |
|
17 |
+
# Alpaca-Llama-3-8B
|
18 |
+
|
19 |
+
|
20 |
+
- **Fine Tuned using dataset:** https://huggingface.co/datasets/yahma/alpaca-cleaned
|
21 |
+
- **Epoch Count:** 1
|
22 |
+
- **Step Count:** 6,470/6,470
|
23 |
+
- **Batch Size:** 2
|
24 |
+
- **Gradient Accumulation Steps:** 4
|
25 |
+
- **Context Size:** 8192
|
26 |
+
- **Num examples:** 51,760
|
27 |
+
- **Trainable Parameters:** 41,943,040
|
28 |
+
- **Learning Rate:** 0.00001
|
29 |
+
- **Training Loss:** 0.960000
|
30 |
+
- **Fined Tuned using:** Google Colab Pro (Nvidia T4 runtime)
|
31 |
|
32 |
- **Developed by:** akumaburn
|
33 |
- **License:** apache-2.0
|
34 |
- **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
|
35 |
+
- **Prompt Format:** Alpaca System (https://libertai.io/apis/text-generation/prompting.html)
|
36 |
+
|
37 |
+
Full model can be found in https://huggingface.co/akumaburn/Alpaca-Llama-3-8B
|
38 |
+
|
39 |
+
mistral-7b-openorca.Q8_0.gguf:
|
40 |
+
- **MMLU-Test:** Final result: **41.5836 +/- 0.4174**
|
41 |
+
- **Arc-Easy:** Final result: 72.6316 +/- 1.8691
|
42 |
+
- **Truthful QA:** Final result: **32.0685 +/- 1.6339**
|
43 |
+
- **Arc-Challenge:** Final result: 48.8294 +/- 2.8956
|
44 |
+
|
45 |
+
llama-3-8b-bnb-4bit.Q8_0.gguf:
|
46 |
+
- **MMLU-Test:** Final result: 40.4074 +/- 0.4156
|
47 |
+
- **Arc-Easy:** Final result: 73.8596 +/- 1.8421
|
48 |
+
- **Truthful QA:** Final result: 26.6830 +/- 1.5484
|
49 |
+
- **Arc-Challenge:** Final result: 46.8227 +/- 2.8906
|
50 |
+
|
51 |
+
Open_Orca_Llama-3-8B-unsloth.Q8_0.gguf:
|
52 |
+
- **MMLU-Test:** Final result: 39.3818 +/- 0.4138
|
53 |
+
- **Arc-Easy:** Final result: 67.3684 +/- 1.9656
|
54 |
+
- **Truthful QA:** Final result: 29.0086 +/- 1.5886
|
55 |
+
- **Arc-Challenge:** Final result: 42.1405 +/- 2.8604
|
56 |
+
|
57 |
+
**Alpaca-Llama-3-8B-GGUF-unsloth.Q8_0.gguf**:
|
58 |
+
- **MMLU-Test:** Final result: 40.6441 +/- 0.4160
|
59 |
+
- **Arc-Easy:** Final result: **77.5439 +/- 1.7494**
|
60 |
+
- **Truthful QA:** Final result: 29.7430 +/- 1.6003
|
61 |
+
- **Arc-Challenge:** Final result: **50.5017 +/- 2.8963**
|
62 |
+
|
63 |
+
Meta-Llama-3-8B.Q8_0.gguf:
|
64 |
+
- **MMLU-Test:** Final result: 40.8664 +/- 0.4163
|
65 |
+
- **Arc-Easy:** Final result: 74.3860 +/- 1.8299
|
66 |
+
- **Truthful QA:** Final result: 28.6414 +/- 1.5826
|
67 |
+
- **Arc-Challenge:** Final result: 47.1572 +/- 2.8917
|
68 |
+
|
69 |
+
Llama.cpp Options For Testing:
|
70 |
+
--samplers "tfs;typical;temp" --draft 32 --ctx-size 8192 --temp 0.82 --tfs 0.8 --typical 1.1 --repeat-last-n 512 --batch-size 8192 --repeat-penalty 1.0 --n-gpu-layers 100 --threads 12
|
71 |
|
72 |
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
|
73 |
|