Update README.md
Browse files
README.md
CHANGED
@@ -14,11 +14,11 @@ language:
|
|
14 |
|
15 |
This is a layer pruning experiment based off of the original llama-3-8b:
|
16 |
|
17 |
-
- 8 layers pruned with PruneMe/MergeKit
|
18 |
- layers selected using [BEE-spoke-data/fineweb-100k_en-med](https://hf.co/datasets/BEE-spoke-data/fineweb-100k_en-med)
|
19 |
- brief subsequent continued pretraining @ ctx 4096
|
20 |
- data: 10k rows of FineWeb (different than pruning data) + some curated data
|
21 |
-
|
22 |
|
23 |
## quick eval
|
24 |
|
|
|
14 |
|
15 |
This is a layer pruning experiment based off of the original llama-3-8b:
|
16 |
|
17 |
+
- 8 layers pruned with [PruneMe](https://github.com/pszemraj/PruneMe/tree/upgrades)/MergeKit
|
18 |
- layers selected using [BEE-spoke-data/fineweb-100k_en-med](https://hf.co/datasets/BEE-spoke-data/fineweb-100k_en-med)
|
19 |
- brief subsequent continued pretraining @ ctx 4096
|
20 |
- data: 10k rows of FineWeb (different than pruning data) + some curated data
|
21 |
+
- wandb [here](https://wandb.ai/pszemraj/llama3-pruning)
|
22 |
|
23 |
## quick eval
|
24 |
|