Upload folder using huggingface_hub
Browse files- config.json +1 -0
- layers.0.mlp/cfg.json +1 -0
- layers.0.mlp/sae.safetensors +3 -0
- layers.12.mlp/cfg.json +1 -0
- layers.12.mlp/sae.safetensors +3 -0
- layers.15.mlp/cfg.json +1 -0
- layers.15.mlp/sae.safetensors +3 -0
- layers.18.mlp/cfg.json +1 -0
- layers.18.mlp/sae.safetensors +3 -0
- layers.21.mlp/cfg.json +1 -0
- layers.21.mlp/sae.safetensors +3 -0
- layers.24.mlp/cfg.json +1 -0
- layers.24.mlp/sae.safetensors +3 -0
- layers.27.mlp/cfg.json +1 -0
- layers.27.mlp/sae.safetensors +3 -0
- layers.3.mlp/cfg.json +1 -0
- layers.3.mlp/sae.safetensors +3 -0
- layers.6.mlp/cfg.json +1 -0
- layers.6.mlp/sae.safetensors +3 -0
- layers.9.mlp/cfg.json +1 -0
- layers.9.mlp/sae.safetensors +3 -0
config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"sae": {"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false}, "batch_size": 8, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["layers.0.mlp", "layers.3.mlp", "layers.6.mlp", "layers.9.mlp", "layers.12.mlp", "layers.15.mlp", "layers.18.mlp", "layers.21.mlp", "layers.24.mlp", "layers.27.mlp"], "init_seeds": [0], "layers": [], "layer_stride": 3, "transcode": false, "distribute_modules": false, "save_every": 1000, "log_to_wandb": true, "run_name": "random-smollm", "wandb_log_frequency": 1, "model": "HuggingFaceTB/SmolLM2-135M", "dataset": "EleutherAI/fineweb-edu-dedup-10b", "split": "train", "ctx_len": 2048, "hf_token": null, "random_init": true, "revision": null, "load_in_8bit": false, "max_examples": null, "resume": false, "text_column": "text", "finetune": null, "shuffle_seed": 42, "data_preprocessing_num_proc": 48}
|
layers.0.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
|
layers.0.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a5bda583ba4e236712a4d431d8826d7bdd10a2a13c5fe2d8a74fb1be2a630602
|
3 |
+
size 170019400
|
layers.12.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
|
layers.12.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3a77476a76a30b087d0b9d42028330fca79a06728d8995a6f6ea4569d9b25e2b
|
3 |
+
size 170019400
|
layers.15.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
|
layers.15.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96c14054ca82305f9d6d2dcc5b79513ab57c6b12edd0aff58bad775318ebe659
|
3 |
+
size 170019400
|
layers.18.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
|
layers.18.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:25d25d7729414704e73c486f92e6ce60af2d62bf36ac6545f31476ddf32335bd
|
3 |
+
size 170019400
|
layers.21.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
|
layers.21.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f54195bbca8f76ee63f2ea7ed182e2f14684b6825f6d8b4c3120ef2cda393d86
|
3 |
+
size 170019400
|
layers.24.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
|
layers.24.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e55a8e095f891ccf197f3283ce900b9420ff9006df595219c2093b67060bbb7
|
3 |
+
size 170019400
|
layers.27.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
|
layers.27.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:336dadf204d17ea9e2798cb3030b961b71dc3071a0a79c7887edeb7e9125fe6d
|
3 |
+
size 170019400
|
layers.3.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
|
layers.3.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2b35dd1ceb12231ea0791c85bfa2a96f209599597b2f2e6224146eb71ec89bb7
|
3 |
+
size 170019400
|
layers.6.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
|
layers.6.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4782288b3f614a0791d6b1bf093162a0e3a1695a881c743f44620eab3ef58c49
|
3 |
+
size 170019400
|
layers.9.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 32, "normalize_decoder": true, "num_latents": 36864, "k": 32, "multi_topk": false, "skip_connection": false, "d_in": 576}
|
layers.9.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb14a20b0f2aecdddee03492eae18506632016f2aaa0aff37696a8cdbf2b6f94
|
3 |
+
size 170019400
|