NeelNanda commited on
Commit
deceebd
1 Parent(s): 6bcc245

Auto Commit

Browse files
gelu-2l_L0_16384_mlp_out_51.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5207a381999b3bcdc99fafd798a1fd8981c5dd4aebb4e41f6acf2e67f8703d9f
3
+ size 67178216
gelu-2l_L0_16384_mlp_out_51_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 51, "batch_size": 4096, "buffer_mult": 384, "lr": 0.0001, "num_tokens": 2000000000, "l1_coeff": 0.0003, "beta1": 0.9, "beta2": 0.99, "dict_mult": 32, "seq_len": 128, "enc_dtype": "fp32", "remove_rare_dir": false, "model_name": "gelu-2l", "site": "mlp_out", "layer": 0, "device": "cuda:1", "model_batch_size": 512, "buffer_size": 1572864, "buffer_batches": 12288, "act_name": "blocks.0.hook_mlp_out", "act_size": 512, "dict_size": 16384}
gelu-2l_L1_16384_mlp_out_50.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cb50ef60320116595a24ed97ef0079c5312bf1445be69641888595308bb4b0f
3
+ size 67178216
gelu-2l_L1_16384_mlp_out_50_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 50, "batch_size": 4096, "buffer_mult": 384, "lr": 0.0001, "num_tokens": 2000000000, "l1_coeff": 0.0003, "beta1": 0.9, "beta2": 0.99, "dict_mult": 32, "seq_len": 128, "enc_dtype": "fp32", "remove_rare_dir": false, "model_name": "gelu-2l", "site": "mlp_out", "layer": 1, "device": "cuda:0", "model_batch_size": 512, "buffer_size": 1572864, "buffer_batches": 12288, "act_name": "blocks.1.hook_mlp_out", "act_size": 512, "dict_size": 16384}