Add files using upload-large-folder tool
Browse files- config.json +1 -0
- layers.0/cfg.json +1 -0
- layers.0/sae.safetensors +3 -0
- layers.1/cfg.json +1 -0
- layers.1/sae.safetensors +3 -0
- layers.10/cfg.json +1 -0
- layers.10/sae.safetensors +3 -0
- layers.11/cfg.json +1 -0
- layers.11/sae.safetensors +3 -0
- layers.12/cfg.json +1 -0
- layers.12/sae.safetensors +3 -0
- layers.13/cfg.json +1 -0
- layers.13/sae.safetensors +3 -0
- layers.14/cfg.json +1 -0
- layers.14/sae.safetensors +3 -0
- layers.15/cfg.json +1 -0
- layers.15/sae.safetensors +3 -0
- layers.16/cfg.json +1 -0
- layers.16/sae.safetensors +3 -0
- layers.17/cfg.json +1 -0
- layers.17/sae.safetensors +3 -0
- layers.18/cfg.json +1 -0
- layers.18/sae.safetensors +3 -0
- layers.19/cfg.json +1 -0
- layers.19/sae.safetensors +3 -0
- layers.2/cfg.json +1 -0
- layers.2/sae.safetensors +3 -0
- layers.20/cfg.json +1 -0
- layers.20/sae.safetensors +3 -0
- layers.21/cfg.json +1 -0
- layers.21/sae.safetensors +3 -0
- layers.22/cfg.json +1 -0
- layers.22/sae.safetensors +3 -0
- layers.23/cfg.json +1 -0
- layers.23/sae.safetensors +3 -0
- layers.3/cfg.json +1 -0
- layers.3/sae.safetensors +3 -0
- layers.4/cfg.json +1 -0
- layers.4/sae.safetensors +3 -0
- layers.5/cfg.json +1 -0
- layers.5/sae.safetensors +3 -0
- layers.6/cfg.json +1 -0
- layers.6/sae.safetensors +3 -0
- layers.7/cfg.json +1 -0
- layers.7/sae.safetensors +3 -0
- layers.8/cfg.json +1 -0
- layers.8/sae.safetensors +3 -0
- layers.9/cfg.json +1 -0
- layers.9/sae.safetensors +3 -0
config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"sae": {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true}, "batch_size": 16, "grad_acc_steps": 1, "micro_acc_steps": 1, "loss_fn": "fvu", "optimizer": "signum", "lr": null, "lr_warmup_steps": 1000, "k_decay_steps": 0, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["layers.0", "layers.1", "layers.2", "layers.3", "layers.4", "layers.5", "layers.6", "layers.7", "layers.8", "layers.9", "layers.10", "layers.11", "layers.12", "layers.13", "layers.14", "layers.15", "layers.16", "layers.17", "layers.18", "layers.19", "layers.20", "layers.21", "layers.22", "layers.23"], "init_seeds": [0], "layers": [0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23], "layer_stride": 1, "distribute_modules": false, "save_every": 10000, "save_best": false, "finetune": null, "log_to_wandb": true, "run_name": null, "wandb_log_frequency": 1, "save_dir": "checkpoints-410m"}
|
layers.0/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.0/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:05afd733f80068bbb4ba78596fe675dcd6f5bf4165349c7b7351826137c1c601
|
3 |
+
size 268570960
|
layers.1/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.1/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01807ce466bad95a773e20f95aaf2588d9a0e3708173834dc72b2149eceae434
|
3 |
+
size 268570960
|
layers.10/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.10/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc03283f8eb349dd576cabd0f980cc94c1a0aa299f4c2429b71175933708fbdd
|
3 |
+
size 268570960
|
layers.11/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.11/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4266572405716cd3d8c4c47b73d45c3e16aee868ff0af145b4230eb39181a263
|
3 |
+
size 268570960
|
layers.12/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.12/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e5621c832dcfce27508051292d9cfc5006bbad1b550283b5323036a75bb0521f
|
3 |
+
size 268570960
|
layers.13/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.13/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e257d64d75cf7dfe0dfb31a312d23f5afd781c965821ca35722efa037bce3a66
|
3 |
+
size 268570960
|
layers.14/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.14/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7eb58e60c1719f3ca2a5f4ba849d3c897dc9abb841467c990b74fd4e1370ac69
|
3 |
+
size 268570960
|
layers.15/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.15/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f033c8fa8c25b84033edb1fabc98ec4fba54649b06ed00a34924c760e245bd99
|
3 |
+
size 268570960
|
layers.16/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.16/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e898991d85776da5a16040b01e770345487efa8e9bbc1c56ba2f2a4ecf11a103
|
3 |
+
size 268570960
|
layers.17/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.17/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:845c40fbec6e16bb2daa4a152c506b8044a0877eb06bd6c4d66a1aae21fbce83
|
3 |
+
size 268570960
|
layers.18/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.18/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37576251eff8820215c965190558255a814f3050ee05d54802a996d9d9e3578e
|
3 |
+
size 268570960
|
layers.19/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.19/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:41015c494124aef9b2fe8ee1daf3730a0355ce988e6091a12b0ec81de8223753
|
3 |
+
size 268570960
|
layers.2/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.2/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ee3c78b689edca3c8edf853c2f132e1e2c9713d7eb6a960a664f270a985d7a4
|
3 |
+
size 268570960
|
layers.20/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.20/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dde88fd647c96fcac66c3ff1954e7e9e20fe6deb2c166bde9905a3911348b114
|
3 |
+
size 268570960
|
layers.21/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.21/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ac9fc40404c2d782fa6d6433741d3d9c9cdd70bf2cf985f310ef907e2572e782
|
3 |
+
size 268570960
|
layers.22/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.22/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e554700c96101bfe8c257244e478c79b23fb445e84b7df00e236e941b8a222d
|
3 |
+
size 268570960
|
layers.23/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.23/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73d888ce7ac3377c239c48b976a8e9b38e027c4ef8fbda052e20b447f3a3454c
|
3 |
+
size 268570960
|
layers.3/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.3/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9b1c2eb62f8b6cc342318b1e91397d570e09fa606eb591b78ac5f2f35dcd652a
|
3 |
+
size 268570960
|
layers.4/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.4/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c580fd0bff9650b5a37287664ef1a2c2cd70205acfbf38057b3ebd60656ddcc
|
3 |
+
size 268570960
|
layers.5/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.5/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c5f4eb7f0b179c42dac39aeb14ba5f8ca51771660be6258eb94f0983ebe65b15
|
3 |
+
size 268570960
|
layers.6/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.6/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c718214b9be8a95f7fe99bdfedf7e5ff6af00427d8de95fc0ec5bdde01910b0c
|
3 |
+
size 268570960
|
layers.7/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.7/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6482ef50e71d69528422768b29db79f3366b1de25da0cf9c51bb12f31de715b0
|
3 |
+
size 268570960
|
layers.8/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.8/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ec57767af44cf9b23071b7c68beb9c21d0af27b06ec55e73528c419a6776e036
|
3 |
+
size 268570960
|
layers.9/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": false, "skip_connection": false, "transcode": true, "d_in": 1024}
|
layers.9/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a6e32e7227fdb06ee82814ed93daec70bc1c0a0d04ceebd78ca486f1d2d29b2e
|
3 |
+
size 268570960
|