pszemraj commited on
Commit
851c3cd
1 Parent(s): da2bc52

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -14,11 +14,11 @@ language:
14
 
15
  This is a layer pruning experiment based off of the original llama-3-8b:
16
 
17
- - 8 layers pruned with PruneMe/MergeKit
18
  - layers selected using [BEE-spoke-data/fineweb-100k_en-med](https://hf.co/datasets/BEE-spoke-data/fineweb-100k_en-med)
19
  - brief subsequent continued pretraining @ ctx 4096
20
  - data: 10k rows of FineWeb (different than pruning data) + some curated data
21
-
22
 
23
  ## quick eval
24
 
 
14
 
15
  This is a layer pruning experiment based off of the original llama-3-8b:
16
 
17
+ - 8 layers pruned with [PruneMe](https://github.com/pszemraj/PruneMe/tree/upgrades)/MergeKit
18
  - layers selected using [BEE-spoke-data/fineweb-100k_en-med](https://hf.co/datasets/BEE-spoke-data/fineweb-100k_en-med)
19
  - brief subsequent continued pretraining @ ctx 4096
20
  - data: 10k rows of FineWeb (different than pruning data) + some curated data
21
+ - wandb [here](https://wandb.ai/pszemraj/llama3-pruning)
22
 
23
  ## quick eval
24