ctigges commited on
Commit
f26d922
·
verified ·
1 Parent(s): 16fc4a0

Upload folder using huggingface_hub

Browse files
0-att-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.0.hook_attn_out", "hook_layer": 0, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
0-att-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed3ab331df675a860677ca80a7a5955c567405e701e38b7fcc0ec5774fe5cb7c
3
+ size 134351168
0-att-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:582aa061a2b9c509f825884534c4afacac3c2918971140d4be2bf6227c71c3e8
3
+ size 131152
1-att-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.1.hook_attn_out", "hook_layer": 1, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
1-att-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94e88d7006cfba7d227f40e0010b631e31562b17977794f76a2d0517ea4727f1
3
+ size 134351168
1-att-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:609ebaf476287d212e7cb09c29700e94427cba68d0b8b60c9f3ee331e007bcd3
3
+ size 131152
2-att-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.2.hook_attn_out", "hook_layer": 2, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
2-att-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c47e46766e13eecb2f92c05d1c341eda05ebf96b5401948326251fcfae999596
3
+ size 134351168
2-att-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04415bf3afed7a496e04d10f688afab3ec537ad5ab6b6f38bdc1fda30bf839c1
3
+ size 131152
3-att-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.3.hook_attn_out", "hook_layer": 3, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
3-att-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0fc0874885822b96eb1b889e65ae9da16d93de65c4cbf2d62fe2a54bb995d61
3
+ size 134351168
3-att-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35564e5bf427ab9b3ba4573058aa4b90290d007baa7d094e5dffbfac7fb80f6f
3
+ size 131152
4-att-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.4.hook_attn_out", "hook_layer": 4, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
4-att-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8763e50063868b4ad274e8b16e9c1d1f6fa749449bdd42fa7514d533fe426b5e
3
+ size 134351168
4-att-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc26e64a0bd01ee857b7a0b08d631f266b7bce6fdf9a50f93dddb1773a9d08f8
3
+ size 131152
5-att-sm/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 512, "d_sae": 32768, "dtype": "torch.float32", "device": "cuda", "model_name": "pythia-70m-deduped", "hook_name": "blocks.5.hook_attn_out", "hook_layer": 5, "hook_head_index": null, "activation_fn_str": "relu", "activation_fn_kwargs": {}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "EleutherAI/the_pile_deduplicated", "dataset_trust_remote_code": true, "context_size": 128, "normalize_activations": "none"}
5-att-sm/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:966aefb665862903965abdba6a38acbb66725191bf4d9d0f0ca7e6ce8a8258ac
3
+ size 134351168
5-att-sm/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:877dbdf1dfa4db34840d64248950928b42e6b5bcc9a42b433ad327548aa624d5
3
+ size 131152