Upload folder using huggingface_hub
Browse files- config.json +1 -0
- layers.10/cfg.json +1 -0
- layers.10/sae.safetensors +3 -0
- layers.12/cfg.json +1 -0
- layers.12/sae.safetensors +3 -0
- layers.14/cfg.json +1 -0
- layers.14/sae.safetensors +3 -0
- layers.16/cfg.json +1 -0
- layers.16/sae.safetensors +3 -0
- layers.18/cfg.json +1 -0
- layers.18/sae.safetensors +3 -0
- layers.20/cfg.json +1 -0
- layers.20/sae.safetensors +3 -0
- layers.22/cfg.json +1 -0
- layers.22/sae.safetensors +3 -0
- layers.6/cfg.json +1 -0
- layers.6/sae.safetensors +3 -0
- layers.8/cfg.json +1 -0
- layers.8/sae.safetensors +3 -0
- lr_scheduler.pt +3 -0
config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"sae": {"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 192, "multi_topk": false}, "batch_size": 24, "grad_acc_steps": 1, "micro_acc_steps": 1, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["layers.6", "layers.8", "layers.10", "layers.12", "layers.14", "layers.16", "layers.18", "layers.20", "layers.22"], "layers": [6, 8, 10, 12, 14, 16, 18, 20, 22], "layer_stride": 1, "distribute_modules": false, "save_every": 1000, "log_to_wandb": true, "run_name": null, "wandb_log_frequency": 1, "model": "Qwen/Qwen2.5-Coder-7B-Instruct", "dataset": "nguyentruong-ins/codeforces_cpp_cleaned_scaled_class", "split": "train", "ctx_len": 756, "hf_token": null, "load_in_8bit": false, "max_examples": null, "resume": false, "seed": 42, "data_preprocessing_num_proc": 24}
|
layers.10/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 192, "multi_topk": false, "d_in": 3584}
|
layers.10/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c7ce8924c6a74598ebae56d661bd6adcf54066eecaf5aed5e64f366780dd9f3e
|
3 |
+
size 822212944
|
layers.12/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 192, "multi_topk": false, "d_in": 3584}
|
layers.12/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c455a4528bad3779ec3b4ce43b4f95c00680a0e7c7555acde23b16e137215fde
|
3 |
+
size 822212944
|
layers.14/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 192, "multi_topk": false, "d_in": 3584}
|
layers.14/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8d98429e2157a2bbcdd57e159b876668e4b3299bd14390cc2896e42196c9683e
|
3 |
+
size 822212944
|
layers.16/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 192, "multi_topk": false, "d_in": 3584}
|
layers.16/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:05056df144b14bd4d548370cb6bcbd8795b3ce940ea7e4e6e61722f3bb7c9240
|
3 |
+
size 822212944
|
layers.18/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 192, "multi_topk": false, "d_in": 3584}
|
layers.18/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c6300708b5bf607f5a35692ee21ba8ac8d2c7636d156135fa13f45f3fc9b3d8b
|
3 |
+
size 822212944
|
layers.20/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 192, "multi_topk": false, "d_in": 3584}
|
layers.20/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9cddd24103985c767b437962aebc39cbd1159e33356b867fb2d7d9c9c27a2da0
|
3 |
+
size 822212944
|
layers.22/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 192, "multi_topk": false, "d_in": 3584}
|
layers.22/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74d118b101ab0fe769159cf99f8dfa58d3c99d54f39e2f071598f8c6e91c33d9
|
3 |
+
size 822212944
|
layers.6/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 192, "multi_topk": false, "d_in": 3584}
|
layers.6/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:609bb33490437418b885223f3dff55a374c641d3d716d5a98df073701610f27f
|
3 |
+
size 822212944
|
layers.8/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"expansion_factor": 8, "normalize_decoder": true, "num_latents": 0, "k": 192, "multi_topk": false, "d_in": 3584}
|
layers.8/sae.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ddc5edc2173ab1c9b2f84a0e6dfa5898124dd79a9cf9eb6ed792eff23378c3e9
|
3 |
+
size 822212944
|
lr_scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bdf0d6ae182233df9b7a948c03c208ed41f3bc227f8be09273441f13f26fa965
|
3 |
+
size 1204
|