Upload folder using huggingface_hub
Browse files- README.md +7 -0
- config.json +65 -0
- n_images_1000079.pt +3 -0
- n_images_1000079_log_feature_sparsity.pt +3 -0
README.md
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Layer: 9
|
2 |
+
Hook: hook_resid_post
|
3 |
+
|
4 |
+
Explained Variance: 0.8374372124671936
|
5 |
+
L0: 291.9156799316406
|
6 |
+
|
7 |
+
Link to Weights & Biases: https://wandb.ai/lolorenz/lorenz_clip_b_l0/runs/8gxnhnb4
|
config.json
ADDED
@@ -0,0 +1,65 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"model_class_name": "HookedViT",
|
3 |
+
"model_name": "open-clip:laion/CLIP-ViT-B-32-DataComp.XL-s13B-b90K",
|
4 |
+
"hook_point": "blocks.9.hook_resid_post",
|
5 |
+
"hook_point_layer": 9,
|
6 |
+
"layer_subtype": "hook_resid_post",
|
7 |
+
"hook_point_head_index": null,
|
8 |
+
"context_size": 50,
|
9 |
+
"use_cached_activations": false,
|
10 |
+
"use_patches_only": false,
|
11 |
+
"cached_activations_path": "activations/_network_scratch_s_sonia.joseph_datasets_kaggle_datasets/open-clip:laion_CLIP-ViT-B-32-DataComp.XL-s13B-b90K/blocks.9.hook_mlp_out",
|
12 |
+
"d_in": 768,
|
13 |
+
"activation_fn_str": "relu",
|
14 |
+
"activation_fn_kwargs": {},
|
15 |
+
"cls_token_only": false,
|
16 |
+
"max_grad_norm": 1.0,
|
17 |
+
"initialization_method": "encoder_transpose_decoder",
|
18 |
+
"normalize_activations": null,
|
19 |
+
"n_batches_in_buffer": 20,
|
20 |
+
"store_batch_size": 32,
|
21 |
+
"num_workers": 16,
|
22 |
+
"num_epochs": 10,
|
23 |
+
"total_training_images": 13000000,
|
24 |
+
"total_training_tokens": 50000000,
|
25 |
+
"image_size": 224,
|
26 |
+
"device": {
|
27 |
+
"__type__": "torch.device",
|
28 |
+
"value": "cuda"
|
29 |
+
},
|
30 |
+
"seed": 42,
|
31 |
+
"dtype": {
|
32 |
+
"__type__": "torch.dtype",
|
33 |
+
"value": "torch.float32"
|
34 |
+
},
|
35 |
+
"architecture": "gated",
|
36 |
+
"sparsity_loss": "l1",
|
37 |
+
"verbose": false,
|
38 |
+
"b_dec_init_method": "geometric_median",
|
39 |
+
"expansion_factor": 64,
|
40 |
+
"from_pretrained_path": null,
|
41 |
+
"d_sae": 49152,
|
42 |
+
"l1_coefficient": 0.9,
|
43 |
+
"lp_norm": 1,
|
44 |
+
"lr": 0.0002,
|
45 |
+
"lr_scheduler_name": "cosineannealingwarmup",
|
46 |
+
"lr_warm_up_steps": 200,
|
47 |
+
"beta1": 0.9,
|
48 |
+
"beta2": 0.999,
|
49 |
+
"train_batch_size": 4096,
|
50 |
+
"dataset_name": "imagenet1k",
|
51 |
+
"dataset_path": "data/ImageNet-complete/",
|
52 |
+
"dataset_train_path": "data/ImageNet-complete/train",
|
53 |
+
"dataset_val_path": "data/ImageNet-complete/val",
|
54 |
+
"use_ghost_grads": false,
|
55 |
+
"feature_sampling_window": 1000,
|
56 |
+
"dead_feature_window": 5000,
|
57 |
+
"dead_feature_threshold": 1e-08,
|
58 |
+
"log_to_wandb": true,
|
59 |
+
"wandb_project": "lorenz_clip_b_l0",
|
60 |
+
"wandb_entity": null,
|
61 |
+
"wandb_log_frequency": 100,
|
62 |
+
"n_validation_runs": 4,
|
63 |
+
"n_checkpoints": 10,
|
64 |
+
"checkpoint_path": "models/sae/clip_B_gated_l0/1e73fd2e-lorenz_clip_b_l0-lorenz_clip_b_l0"
|
65 |
+
}
|
n_images_1000079.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe3dcaea58ba024014d42215f8527e9543b1b109487642f1845b57830ac8f5fa
|
3 |
+
size 302784743
|
n_images_1000079_log_feature_sparsity.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b3541f72abcf75888d36644e827541020e1f708b947a693337afd843189611a
|
3 |
+
size 198002
|