Update README.md
Browse files
README.md
CHANGED
@@ -1,31 +1,40 @@
|
|
1 |
---
|
2 |
license: unlicense
|
3 |
---
|
|
|
|
|
4 |
This is a testing LORA for Llama-3 8B or Llama-3 8B Instruct. The goal for this model was to bring back some of the expressive prose and writing style of the base model, as well as shift the rather dry style of the 8B instruct.
|
5 |
-
Data Processing
|
6 |
-
Raw Data to Custom Data
|
7 |
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
|
14 |
-
Data Cleaning Code
|
15 |
Once the training is verified to be beneficial (and thus the cleaning was likely to be correct), the data cleaning code will be released.
|
16 |
-
|
17 |
-
Training
|
18 |
-
|
19 |
-
Training
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: unlicense
|
3 |
---
|
4 |
+
## Overview
|
5 |
+
|
6 |
This is a testing LORA for Llama-3 8B or Llama-3 8B Instruct. The goal for this model was to bring back some of the expressive prose and writing style of the base model, as well as shift the rather dry style of the 8B instruct.
|
|
|
|
|
7 |
|
8 |
+
## Data Processing
|
9 |
+
|
10 |
+
### Raw Data to Custom Data
|
11 |
+
|
12 |
+
- Started with ~40GB of raw data
|
13 |
+
- Aggressively selected for writing style
|
14 |
+
- Cleaned multiple times both automatically and by hand
|
15 |
+
- Final dataset size: 78.4MB
|
16 |
+
- No synthetic data present in the used dataset
|
17 |
+
|
18 |
+
### Data Cleaning Code
|
19 |
|
|
|
20 |
Once the training is verified to be beneficial (and thus the cleaning was likely to be correct), the data cleaning code will be released.
|
21 |
+
|
22 |
+
## Training Procedure
|
23 |
+
|
24 |
+
### Training Framework
|
25 |
+
|
26 |
+
Training was done QLORA style via [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl). The full training script along with the data processing scripts will be released similarly once the procedure is verified to benefit the model in a useful way.
|
27 |
+
|
28 |
+
### Training Parameters
|
29 |
+
|
30 |
+
- Base Model: Llama 3 8B (Non instruct)
|
31 |
+
- r: 4
|
32 |
+
- alpha: 8
|
33 |
+
- dropout: 0
|
34 |
+
- warmup: 45 steps
|
35 |
+
- epochs: 2
|
36 |
+
- lr: constant with warmup
|
37 |
+
- optimizer: adamw (torch fused)
|
38 |
+
- weight decay: 0.1
|
39 |
+
- adam_b1: 0.9
|
40 |
+
- adam_b2: 0.999
|