Upload SAE gen_001/blocks.22.hook_resid_post
Browse files
gen_001/blocks.22.hook_resid_post/cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"d_in": 2304, "d_sae": 16384, "dtype": "float32", "device": "cuda:0", "apply_b_dec_to_input": true, "normalize_activations": "none", "reshape_activations": "none", "metadata": {"sae_lens_version": "6.36.2", "sae_lens_training_version": "6.36.2", "dataset_path": "monology/pile-uncopyrighted", "hook_name": "blocks.22.hook_resid_post", "model_name": "gemma-2-2b", "model_class_name": "HookedTransformer", "hook_head_index": null, "context_size": 1024, "seqpos_slice": [null], "model_from_pretrained_kwargs": {"center_writing_weights": false}, "prepend_bos": true, "exclude_special_tokens": false, "sequence_separator_token": "bos", "disable_concat_sequences": false}, "decoder_init_norm": 0.1, "jumprelu_init_threshold": 0.001, "jumprelu_bandwidth": 0.001, "jumprelu_sparsity_loss_mode": "step", "l0_coefficient": 0.001, "l0_warm_up_steps": 10000, "pre_act_loss_coefficient": null, "jumprelu_tanh_scale": 4.0, "architecture": "jumprelu"}
|
gen_001/blocks.22.hook_resid_post/sae_weights.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b2f300ef486e7fdc51983a05fe4740b8cb50e0e57e5876ff4263418c11a71626
|
| 3 |
+
size 302130576
|