norabelrose commited on
Commit
40a22c2
·
verified ·
1 Parent(s): 9684c17

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false}, "batch_size": 8, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["layers.0.mlp", "layers.3.mlp", "layers.6.mlp", "layers.9.mlp", "layers.12.mlp", "layers.15.mlp", "layers.18.mlp", "layers.21.mlp", "layers.24.mlp", "layers.27.mlp"], "init_seeds": [0], "layers": [], "layer_stride": 3, "transcode": false, "distribute_modules": false, "save_every": 1000, "log_to_wandb": true, "run_name": "random-smollm", "wandb_log_frequency": 1, "model": "HuggingFaceTB/SmolLM2-135M", "dataset": "EleutherAI/fineweb-edu-dedup-10b", "split": "train", "ctx_len": 2048, "hf_token": null, "random_init": true, "revision": null, "load_in_8bit": false, "max_examples": null, "resume": false, "text_column": "text", "finetune": null, "shuffle_seed": 42, "data_preprocessing_num_proc": 48}
layers.0.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
layers.0.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5bda583ba4e236712a4d431d8826d7bdd10a2a13c5fe2d8a74fb1be2a630602
3
+ size 170019400
layers.12.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
layers.12.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a77476a76a30b087d0b9d42028330fca79a06728d8995a6f6ea4569d9b25e2b
3
+ size 170019400
layers.15.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
layers.15.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96c14054ca82305f9d6d2dcc5b79513ab57c6b12edd0aff58bad775318ebe659
3
+ size 170019400
layers.18.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
layers.18.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25d25d7729414704e73c486f92e6ce60af2d62bf36ac6545f31476ddf32335bd
3
+ size 170019400
layers.21.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
layers.21.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f54195bbca8f76ee63f2ea7ed182e2f14684b6825f6d8b4c3120ef2cda393d86
3
+ size 170019400
layers.24.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
layers.24.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e55a8e095f891ccf197f3283ce900b9420ff9006df595219c2093b67060bbb7
3
+ size 170019400
layers.27.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
layers.27.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:336dadf204d17ea9e2798cb3030b961b71dc3071a0a79c7887edeb7e9125fe6d
3
+ size 170019400
layers.3.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
layers.3.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b35dd1ceb12231ea0791c85bfa2a96f209599597b2f2e6224146eb71ec89bb7
3
+ size 170019400
layers.6.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
layers.6.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4782288b3f614a0791d6b1bf093162a0e3a1695a881c743f44620eab3ef58c49
3
+ size 170019400
layers.9.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
layers.9.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb14a20b0f2aecdddee03492eae18506632016f2aaa0aff37696a8cdbf2b6f94
3
+ size 170019400