Add files using upload-large-folder tool
Browse files- layers.0.mlp/cfg.json +1 -0
- layers.0.mlp/sae.safetensors +3 -0
- layers.1.mlp/cfg.json +1 -0
- layers.1.mlp/sae.safetensors +3 -0
- layers.10.mlp/cfg.json +1 -0
- layers.10.mlp/sae.safetensors +3 -0
- layers.11.mlp/cfg.json +1 -0
- layers.11.mlp/sae.safetensors +3 -0
- layers.12.mlp/cfg.json +1 -0
- layers.12.mlp/sae.safetensors +3 -0
- layers.13.mlp/cfg.json +1 -0
- layers.13.mlp/sae.safetensors +3 -0
- layers.14.mlp/cfg.json +1 -0
- layers.14.mlp/sae.safetensors +3 -0
- layers.15.mlp/cfg.json +1 -0
- layers.15.mlp/sae.safetensors +3 -0
- layers.2.mlp/cfg.json +1 -0
- layers.2.mlp/sae.safetensors +3 -0
- layers.3.mlp/cfg.json +1 -0
- layers.3.mlp/sae.safetensors +3 -0
- layers.4.mlp/cfg.json +1 -0
- layers.4.mlp/sae.safetensors +3 -0
- layers.5.mlp/cfg.json +1 -0
- layers.5.mlp/sae.safetensors +3 -0
- layers.6.mlp/cfg.json +1 -0
- layers.6.mlp/sae.safetensors +3 -0
- layers.7.mlp/cfg.json +1 -0
- layers.7.mlp/sae.safetensors +3 -0
- layers.8.mlp/cfg.json +1 -0
- layers.8.mlp/sae.safetensors +3 -0
- layers.9.mlp/cfg.json +1 -0
- layers.9.mlp/sae.safetensors +3 -0
layers.0.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.0.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8b640ae0512948ce3f47a9ad92a057bf578c4f62eabf9e508c51bcd0d449476c
|
3 |
+
size 1074008408
|
layers.1.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.1.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:76c594b4e82ea5627d7b6fd72c3e28d8735e46e3e9f258e3d2bfc0e025d63e77
|
3 |
+
size 1074008408
|
layers.10.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.10.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3d8156951aef15e463d31fe66f6352aa6a2a572139aff97277d5025e1a60353
|
3 |
+
size 1074008408
|
layers.11.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.11.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7029a14028a70a05acce5c209e5d428c14c0e430ed00e5f3eaab5a3ee389eb1f
|
3 |
+
size 1074008408
|
layers.12.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.12.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2748ecd2d32247d97effdf53bf5212071a02e2b243202fd79dde3610e9c60904
|
3 |
+
size 1074008408
|
layers.13.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.13.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:91d1d6ea425cae973c3420f4f970f354b897f1f6f31ed1de0fad2d508011983f
|
3 |
+
size 1074008408
|
layers.14.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.14.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b7f29b67d888ddf9f828c23e66035860f873ca898d63eb058dae16864069331c
|
3 |
+
size 1074008408
|
layers.15.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.15.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:caf1600b1b7597898bd1fd6e9c52bb1dd9a017e05e743816499915114303faf0
|
3 |
+
size 1074008408
|
layers.2.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.2.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:39e0f72b0447e163a7141b5311aeb363cb164a689dd0e1f792e1261e6f60fcb3
|
3 |
+
size 1074008408
|
layers.3.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.3.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ebd8d8361d80f4bd7d3f028ed41be4b6db771aecf976fe46cf3947baac6fdd14
|
3 |
+
size 1074008408
|
layers.4.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.4.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:67d27413c25bff3e47e1affc5c19915c41ac423a2fd3cf5184e874a97583d046
|
3 |
+
size 1074008408
|
layers.5.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.5.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:374912589791a27a675de8dc874a5a7e87570d2749c5b2eabd3b3f5a1c69010e
|
3 |
+
size 1074008408
|
layers.6.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.6.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:922e401bca018b57844c3756af46dfabbdae982c70f5caa865891e7262092111
|
3 |
+
size 1074008408
|
layers.7.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.7.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e58a386541eafb4a6dcc5d4f97769b81dca28e9867f81e9a03dc0d292c66312b
|
3 |
+
size 1074008408
|
layers.8.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.8.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a10ab4de7156362e05b69df713444bea3f84809189f26c49d51a0ab4db401a8d
|
3 |
+
size 1074008408
|
layers.9.mlp/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"dtype": "none", "activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 131072, "k": 64, "multi_topk": false, "skip_connection": false, "transcode": true, "tp_output": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "per_source_tied": false, "secondary_target_tied": false, "coalesce_topk": "none", "topk_coalesced": false, "use_fp8": false, "d_in": 2048}
|
layers.9.mlp/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6ec47b3919098e76b8779f085d34e3a28056d5d5f30c79e51c0a1f831eefb972
|
3 |
+
size 1074008408
|