nixgd commited on
Commit
7135151
1 Parent(s): 5dab464
downstream_similar_ce_layer_10.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:071ebefad984390288ff036dc2952ffad150ca9be411be06667cb58b7d19c536
3
+ size 283305160
downstream_similar_ce_layer_10.yaml ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 8
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: null
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 40000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig:
21
+ hook_positions:
22
+ - blocks.11.hook_resid_pre
23
+ total_coeff: 0.05
24
+ logits_kl:
25
+ coeff: 0.5
26
+ out_to_in:
27
+ coeff: 0.0
28
+ out_to_orig: null
29
+ sparsity:
30
+ coeff: 1.75
31
+ p_norm: 1.0
32
+ lr: 0.0005
33
+ lr_schedule: cosine
34
+ max_grad_norm: 10.0
35
+ min_lr_factor: 0.1
36
+ n_samples: 400000
37
+ saes:
38
+ dict_size_to_input_ratio: 60.0
39
+ k: null
40
+ pretrained_sae_paths: null
41
+ retrain_saes: false
42
+ sae_positions:
43
+ - blocks.10.hook_resid_pre
44
+ type_of_sparsifier: sae
45
+ save_dir: /data/dan_braun/sparsify/sparsify/scripts/train_tlens_saes/out
46
+ save_every_n_samples: null
47
+ seed: 0
48
+ tlens_model_name: gpt2-small
49
+ tlens_model_path: null
50
+ train_data:
51
+ column_name: input_ids
52
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
53
+ is_tokenized: true
54
+ n_ctx: 1024
55
+ seed: null
56
+ split: train
57
+ streaming: true
58
+ tokenizer_name: gpt2
59
+ wandb_project: gpt2-e2e
60
+ wandb_run_name: null
61
+ wandb_run_name_prefix: ''
62
+ warmup_samples: 20000
downstream_similar_ce_layer_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbe566a12b302d22db01017fdc0059f2e3479c84654c03c2866623ce73002fdd
3
+ size 283305152
downstream_similar_ce_layer_2.yaml ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 4
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
9
+ is_tokenized: true
10
+ n_ctx: 1024
11
+ split: train
12
+ streaming: true
13
+ tokenizer_name: gpt2
14
+ eval_every_n_samples: 20000
15
+ eval_n_samples: 200
16
+ log_every_n_grad_steps: 20
17
+ loss:
18
+ in_to_orig:
19
+ hook_positions:
20
+ - blocks.3.hook_resid_pre
21
+ - blocks.4.hook_resid_pre
22
+ - blocks.5.hook_resid_pre
23
+ - blocks.6.hook_resid_pre
24
+ - blocks.7.hook_resid_pre
25
+ - blocks.8.hook_resid_pre
26
+ - blocks.9.hook_resid_pre
27
+ - blocks.10.hook_resid_pre
28
+ - blocks.11.hook_resid_pre
29
+ total_coeff: 2.5
30
+ logits_kl:
31
+ coeff: 0.5
32
+ out_to_in:
33
+ coeff: 0
34
+ out_to_orig: null
35
+ sparsity:
36
+ coeff: 10
37
+ lr: 0.0005
38
+ lr_schedule: cosine
39
+ max_grad_norm: 10
40
+ min_lr_factor: 0.1
41
+ n_samples: 400000
42
+ saes:
43
+ dict_size_to_input_ratio: 60
44
+ sae_positions: blocks.2.hook_resid_pre
45
+ save_dir: /data/jordan_tensor/sparsify/sparsify/scripts/train_tlens_saes/out
46
+ save_every_n_samples: null
47
+ seed: 0
48
+ tlens_model_name: gpt2-small
49
+ tlens_model_path: null
50
+ train_data:
51
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
52
+ is_tokenized: true
53
+ n_ctx: 1024
54
+ split: train
55
+ streaming: true
56
+ tokenizer_name: gpt2
57
+ wandb_project: gpt2-e2e
58
+ wandb_run_name: null
59
+ wandb_run_name_prefix: recon_
60
+ warmup_samples: 20000
downstream_similar_ce_layer_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68a2aaad18bf2855fb6086ce68bdf6e54a6911a89918ae86b41fff2daee17712
3
+ size 283305152
downstream_similar_ce_layer_6.yaml ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 4
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: null
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 20000
17
+ eval_n_samples: 200
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig:
21
+ hook_positions:
22
+ - blocks.7.hook_resid_pre
23
+ - blocks.8.hook_resid_pre
24
+ - blocks.9.hook_resid_pre
25
+ - blocks.10.hook_resid_pre
26
+ - blocks.11.hook_resid_pre
27
+ total_coeff: 2.5
28
+ logits_kl:
29
+ coeff: 0.5
30
+ out_to_in:
31
+ coeff: 0.0
32
+ out_to_orig: null
33
+ sparsity:
34
+ coeff: 50.0
35
+ p_norm: 1.0
36
+ lr: 0.0005
37
+ lr_schedule: cosine
38
+ max_grad_norm: 10.0
39
+ min_lr_factor: 0.1
40
+ n_samples: 400000
41
+ saes:
42
+ dict_size_to_input_ratio: 60.0
43
+ k: null
44
+ pretrained_sae_paths: null
45
+ retrain_saes: false
46
+ sae_positions:
47
+ - blocks.6.hook_resid_pre
48
+ type_of_sparsifier: sae
49
+ save_dir: /data/jordan_tensor/sparsify/sparsify/scripts/train_tlens_saes/out
50
+ save_every_n_samples: null
51
+ seed: 0
52
+ tlens_model_name: gpt2-small
53
+ tlens_model_path: null
54
+ train_data:
55
+ column_name: input_ids
56
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
57
+ is_tokenized: true
58
+ n_ctx: 1024
59
+ seed: null
60
+ split: train
61
+ streaming: true
62
+ tokenizer_name: gpt2
63
+ wandb_project: gpt2-e2e
64
+ wandb_run_name: null
65
+ wandb_run_name_prefix: recon_
66
+ warmup_samples: 20000
downstream_similar_l0_layer_10.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cf449615184314b005e2eb0e5378e4ae5cc065f42bc3e748bb988476cabd447
3
+ size 283305160
downstream_similar_l0_layer_10.yaml ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 8
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: null
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 40000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig:
21
+ hook_positions:
22
+ - blocks.11.hook_resid_pre
23
+ total_coeff: 0.05
24
+ logits_kl:
25
+ coeff: 0.5
26
+ out_to_in:
27
+ coeff: 0.0
28
+ out_to_orig: null
29
+ sparsity:
30
+ coeff: 1.75
31
+ p_norm: 1.0
32
+ lr: 0.0005
33
+ lr_schedule: cosine
34
+ max_grad_norm: 10.0
35
+ min_lr_factor: 0.1
36
+ n_samples: 400000
37
+ saes:
38
+ dict_size_to_input_ratio: 60.0
39
+ k: null
40
+ pretrained_sae_paths: null
41
+ retrain_saes: false
42
+ sae_positions:
43
+ - blocks.10.hook_resid_pre
44
+ type_of_sparsifier: sae
45
+ save_dir: /data/dan_braun/sparsify/sparsify/scripts/train_tlens_saes/out
46
+ save_every_n_samples: null
47
+ seed: 0
48
+ tlens_model_name: gpt2-small
49
+ tlens_model_path: null
50
+ train_data:
51
+ column_name: input_ids
52
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
53
+ is_tokenized: true
54
+ n_ctx: 1024
55
+ seed: null
56
+ split: train
57
+ streaming: true
58
+ tokenizer_name: gpt2
59
+ wandb_project: gpt2-e2e
60
+ wandb_run_name: null
61
+ wandb_run_name_prefix: ''
62
+ warmup_samples: 20000
downstream_similar_l0_layer_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99744a3807d0e2287c62c680a56108f5a82a46c0b4b6e02344847b04608a792b
3
+ size 283305152
downstream_similar_l0_layer_2.yaml ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 4
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
9
+ is_tokenized: true
10
+ n_ctx: 1024
11
+ split: train
12
+ streaming: true
13
+ tokenizer_name: gpt2
14
+ eval_every_n_samples: 20000
15
+ eval_n_samples: 200
16
+ log_every_n_grad_steps: 20
17
+ loss:
18
+ in_to_orig:
19
+ hook_positions:
20
+ - blocks.3.hook_resid_pre
21
+ - blocks.4.hook_resid_pre
22
+ - blocks.5.hook_resid_pre
23
+ - blocks.6.hook_resid_pre
24
+ - blocks.7.hook_resid_pre
25
+ - blocks.8.hook_resid_pre
26
+ - blocks.9.hook_resid_pre
27
+ - blocks.10.hook_resid_pre
28
+ - blocks.11.hook_resid_pre
29
+ total_coeff: 2.5
30
+ logits_kl:
31
+ coeff: 0.5
32
+ out_to_in:
33
+ coeff: 0
34
+ out_to_orig: null
35
+ sparsity:
36
+ coeff: 35
37
+ lr: 0.0005
38
+ lr_schedule: cosine
39
+ max_grad_norm: 10
40
+ min_lr_factor: 0.1
41
+ n_samples: 400000
42
+ saes:
43
+ dict_size_to_input_ratio: 60
44
+ sae_positions: blocks.2.hook_resid_pre
45
+ save_dir: /data/jordan_tensor/sparsify/sparsify/scripts/train_tlens_saes/out
46
+ save_every_n_samples: null
47
+ seed: 0
48
+ tlens_model_name: gpt2-small
49
+ tlens_model_path: null
50
+ train_data:
51
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
52
+ is_tokenized: true
53
+ n_ctx: 1024
54
+ split: train
55
+ streaming: true
56
+ tokenizer_name: gpt2
57
+ wandb_project: gpt2-e2e
58
+ wandb_run_name: null
59
+ wandb_run_name_prefix: recon_
60
+ warmup_samples: 20000
downstream_similar_l0_layer_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76e7e12a4666688728574c762a20d92a82c0a6a403f77ef15968349a89548e22
3
+ size 283305152
downstream_similar_l0_layer_6.yaml ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 4
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: null
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 20000
17
+ eval_n_samples: 200
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig:
21
+ hook_positions:
22
+ - blocks.7.hook_resid_pre
23
+ - blocks.8.hook_resid_pre
24
+ - blocks.9.hook_resid_pre
25
+ - blocks.10.hook_resid_pre
26
+ - blocks.11.hook_resid_pre
27
+ total_coeff: 2.5
28
+ logits_kl:
29
+ coeff: 0.5
30
+ out_to_in:
31
+ coeff: 0.0
32
+ out_to_orig: null
33
+ sparsity:
34
+ coeff: 50.0
35
+ p_norm: 1.0
36
+ lr: 0.0005
37
+ lr_schedule: cosine
38
+ max_grad_norm: 10.0
39
+ min_lr_factor: 0.1
40
+ n_samples: 400000
41
+ saes:
42
+ dict_size_to_input_ratio: 60.0
43
+ k: null
44
+ pretrained_sae_paths: null
45
+ retrain_saes: false
46
+ sae_positions:
47
+ - blocks.6.hook_resid_pre
48
+ type_of_sparsifier: sae
49
+ save_dir: /data/jordan_tensor/sparsify/sparsify/scripts/train_tlens_saes/out
50
+ save_every_n_samples: null
51
+ seed: 0
52
+ tlens_model_name: gpt2-small
53
+ tlens_model_path: null
54
+ train_data:
55
+ column_name: input_ids
56
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
57
+ is_tokenized: true
58
+ n_ctx: 1024
59
+ seed: null
60
+ split: train
61
+ streaming: true
62
+ tokenizer_name: gpt2
63
+ wandb_project: gpt2-e2e
64
+ wandb_run_name: null
65
+ wandb_run_name_prefix: recon_
66
+ warmup_samples: 20000
e2e_similar_ce_layer_10.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f98c3f3f7f21431e4025515987c05ad1edf4a68b78935d7748cf81a0486a85cb
3
+ size 283305104
e2e_similar_ce_layer_10.yaml ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 16
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 20000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl:
22
+ coeff: 1.0
23
+ out_to_in:
24
+ coeff: 0.0
25
+ out_to_orig: null
26
+ sparsity:
27
+ coeff: 1.5
28
+ p_norm: 1.0
29
+ lr: 0.0005
30
+ lr_schedule: cosine
31
+ max_grad_norm: 10.0
32
+ min_lr_factor: 0.1
33
+ n_samples: 400000
34
+ saes:
35
+ dict_size_to_input_ratio: 60.0
36
+ k: null
37
+ pretrained_sae_paths: null
38
+ retrain_saes: false
39
+ sae_positions:
40
+ - blocks.10.hook_resid_pre
41
+ type_of_sparsifier: sae
42
+ save_dir: /data/jordan_tensor/sparsify/sparsify/scripts/train_tlens_saes/out
43
+ save_every_n_samples: null
44
+ seed: 0
45
+ tlens_model_name: gpt2-small
46
+ tlens_model_path: null
47
+ train_data:
48
+ column_name: input_ids
49
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
50
+ is_tokenized: true
51
+ n_ctx: 1024
52
+ seed: 0
53
+ split: train
54
+ streaming: true
55
+ tokenizer_name: gpt2
56
+ wandb_project: gpt2-e2e_play
57
+ wandb_run_name: null
58
+ wandb_run_name_prefix: ''
59
+ warmup_samples: 20000
e2e_similar_ce_layer_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4864807ce659ebc967867707687982f6c3c4f3b61f4f880de5427a26bf7001bf
3
+ size 283305096
e2e_similar_ce_layer_2.yaml ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 16
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 20000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl:
22
+ coeff: 1.0
23
+ out_to_in:
24
+ coeff: 0.0
25
+ out_to_orig: null
26
+ sparsity:
27
+ coeff: 0.5
28
+ p_norm: 1.0
29
+ lr: 0.0005
30
+ lr_schedule: cosine
31
+ max_grad_norm: 10.0
32
+ min_lr_factor: 0.1
33
+ n_samples: 400000
34
+ saes:
35
+ dict_size_to_input_ratio: 60.0
36
+ k: null
37
+ pretrained_sae_paths: null
38
+ retrain_saes: false
39
+ sae_positions:
40
+ - blocks.2.hook_resid_pre
41
+ type_of_sparsifier: sae
42
+ save_dir: /data/jordan_tensor/sparsify/sparsify/scripts/train_tlens_saes/out
43
+ save_every_n_samples: null
44
+ seed: 0
45
+ tlens_model_name: gpt2-small
46
+ tlens_model_path: null
47
+ train_data:
48
+ column_name: input_ids
49
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
50
+ is_tokenized: true
51
+ n_ctx: 1024
52
+ seed: 0
53
+ split: train
54
+ streaming: true
55
+ tokenizer_name: gpt2
56
+ wandb_project: gpt2-e2e_play
57
+ wandb_run_name: null
58
+ wandb_run_name_prefix: ''
59
+ warmup_samples: 20000
e2e_similar_ce_layer_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:252358bb0c5bb9a5f18ad8c9e1cf4257948c6661af4f17ca24a162b1edb709b7
3
+ size 283305096
e2e_similar_ce_layer_6.yaml ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 16
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 20000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl:
22
+ coeff: 1.0
23
+ out_to_in:
24
+ coeff: 0.0
25
+ out_to_orig: null
26
+ sparsity:
27
+ coeff: 3.0
28
+ p_norm: 1.0
29
+ lr: 0.0005
30
+ lr_schedule: cosine
31
+ max_grad_norm: 10.0
32
+ min_lr_factor: 0.1
33
+ n_samples: 400000
34
+ saes:
35
+ dict_size_to_input_ratio: 60.0
36
+ k: null
37
+ pretrained_sae_paths: null
38
+ retrain_saes: false
39
+ sae_positions:
40
+ - blocks.6.hook_resid_pre
41
+ type_of_sparsifier: sae
42
+ save_dir: /data/jordan_tensor/sparsify/sparsify/scripts/train_tlens_saes/out
43
+ save_every_n_samples: null
44
+ seed: 0
45
+ tlens_model_name: gpt2-small
46
+ tlens_model_path: null
47
+ train_data:
48
+ column_name: input_ids
49
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
50
+ is_tokenized: true
51
+ n_ctx: 1024
52
+ seed: 0
53
+ split: train
54
+ streaming: true
55
+ tokenizer_name: gpt2
56
+ wandb_project: gpt2-e2e_play
57
+ wandb_run_name: null
58
+ wandb_run_name_prefix: ''
59
+ warmup_samples: 20000
e2e_similar_l0_layer_10.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d8369abe8f4e8962eff009b9fb7f681a046422dce32d01e1fe99725ab6eb86b
3
+ size 283305104
e2e_similar_l0_layer_10.yaml ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 16
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 20000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl:
22
+ coeff: 1.0
23
+ out_to_in:
24
+ coeff: 0.0
25
+ out_to_orig: null
26
+ sparsity:
27
+ coeff: 1.5
28
+ p_norm: 1.0
29
+ lr: 0.0005
30
+ lr_schedule: cosine
31
+ max_grad_norm: 10.0
32
+ min_lr_factor: 0.1
33
+ n_samples: 400000
34
+ saes:
35
+ dict_size_to_input_ratio: 60.0
36
+ k: null
37
+ pretrained_sae_paths: null
38
+ retrain_saes: false
39
+ sae_positions:
40
+ - blocks.10.hook_resid_pre
41
+ type_of_sparsifier: sae
42
+ save_dir: /data/jordan_tensor/sparsify/sparsify/scripts/train_tlens_saes/out
43
+ save_every_n_samples: null
44
+ seed: 0
45
+ tlens_model_name: gpt2-small
46
+ tlens_model_path: null
47
+ train_data:
48
+ column_name: input_ids
49
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
50
+ is_tokenized: true
51
+ n_ctx: 1024
52
+ seed: 0
53
+ split: train
54
+ streaming: true
55
+ tokenizer_name: gpt2
56
+ wandb_project: gpt2-e2e_play
57
+ wandb_run_name: null
58
+ wandb_run_name_prefix: ''
59
+ warmup_samples: 20000
e2e_similar_l0_layer_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44a7b30cff15bfac66eb46361627016ba1a6549ae15c8740c3119e33fcfd0c2d
3
+ size 283305096
e2e_similar_l0_layer_2.yaml ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 16
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 20000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl:
22
+ coeff: 1.0
23
+ out_to_in:
24
+ coeff: 0.0
25
+ out_to_orig: null
26
+ sparsity:
27
+ coeff: 1.5
28
+ p_norm: 1.0
29
+ lr: 0.0005
30
+ lr_schedule: cosine
31
+ max_grad_norm: 10.0
32
+ min_lr_factor: 0.1
33
+ n_samples: 400000
34
+ saes:
35
+ dict_size_to_input_ratio: 60.0
36
+ k: null
37
+ pretrained_sae_paths: null
38
+ retrain_saes: false
39
+ sae_positions:
40
+ - blocks.2.hook_resid_pre
41
+ type_of_sparsifier: sae
42
+ save_dir: /data/jordan_tensor/sparsify/sparsify/scripts/train_tlens_saes/out
43
+ save_every_n_samples: null
44
+ seed: 0
45
+ tlens_model_name: gpt2-small
46
+ tlens_model_path: null
47
+ train_data:
48
+ column_name: input_ids
49
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
50
+ is_tokenized: true
51
+ n_ctx: 1024
52
+ seed: 0
53
+ split: train
54
+ streaming: true
55
+ tokenizer_name: gpt2
56
+ wandb_project: gpt2-e2e_play
57
+ wandb_run_name: null
58
+ wandb_run_name_prefix: ''
59
+ warmup_samples: 20000
e2e_similar_l0_layer_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf5513c22c99845afcaf125ff203fdfac9d49477f99c59c0912aefe6c965d33f
3
+ size 283305096
e2e_similar_l0_layer_6.yaml ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 16
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 20000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl:
22
+ coeff: 1.0
23
+ out_to_in:
24
+ coeff: 0.0
25
+ out_to_orig: null
26
+ sparsity:
27
+ coeff: 1.5
28
+ p_norm: 1.0
29
+ lr: 0.0005
30
+ lr_schedule: cosine
31
+ max_grad_norm: 10.0
32
+ min_lr_factor: 0.1
33
+ n_samples: 400000
34
+ saes:
35
+ dict_size_to_input_ratio: 60.0
36
+ k: null
37
+ pretrained_sae_paths: null
38
+ retrain_saes: false
39
+ sae_positions:
40
+ - blocks.6.hook_resid_pre
41
+ type_of_sparsifier: sae
42
+ save_dir: /data/jordan_tensor/sparsify/sparsify/scripts/train_tlens_saes/out
43
+ save_every_n_samples: null
44
+ seed: 0
45
+ tlens_model_name: gpt2-small
46
+ tlens_model_path: null
47
+ train_data:
48
+ column_name: input_ids
49
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
50
+ is_tokenized: true
51
+ n_ctx: 1024
52
+ seed: 0
53
+ split: train
54
+ streaming: true
55
+ tokenizer_name: gpt2
56
+ wandb_project: gpt2-e2e_play
57
+ wandb_run_name: null
58
+ wandb_run_name_prefix: ''
59
+ warmup_samples: 20000
local_similar_ce_layer_10.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:353e4397a1910189a7317c5882804226a884fc5cf3bc03aa4b55bbbadfdaf161
3
+ size 283305120
local_similar_ce_layer_10.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 8
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 40000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl: null
22
+ out_to_in:
23
+ coeff: 1.0
24
+ out_to_orig: null
25
+ sparsity:
26
+ coeff: 6.0
27
+ p_norm: 1.0
28
+ lr: 0.0005
29
+ lr_schedule: cosine
30
+ max_grad_norm: 10.0
31
+ min_lr_factor: 0.1
32
+ n_samples: 400000
33
+ saes:
34
+ dict_size_to_input_ratio: 60.0
35
+ k: null
36
+ pretrained_sae_paths: null
37
+ retrain_saes: false
38
+ sae_positions:
39
+ - blocks.10.hook_resid_pre
40
+ type_of_sparsifier: sae
41
+ save_dir: /mnt/ssd-interp/dan/sparsify/sparsify/scripts/train_tlens_saes/out
42
+ save_every_n_samples: null
43
+ seed: 0
44
+ tlens_model_name: gpt2-small
45
+ tlens_model_path: null
46
+ train_data:
47
+ column_name: input_ids
48
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
49
+ is_tokenized: true
50
+ n_ctx: 1024
51
+ seed: 0
52
+ split: train
53
+ streaming: true
54
+ tokenizer_name: gpt2
55
+ wandb_project: gpt2-layerwise_play
56
+ wandb_run_name: null
57
+ wandb_run_name_prefix: ''
58
+ warmup_samples: 20000
local_similar_ce_layer_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49e197d179193f62de6633061aa82b74b2b8bec78a6844ceed8195f9539686c0
3
+ size 283305112
local_similar_ce_layer_2.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 8
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 40000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl: null
22
+ out_to_in:
23
+ coeff: 1.0
24
+ out_to_orig: null
25
+ sparsity:
26
+ coeff: 0.8
27
+ p_norm: 1.0
28
+ lr: 0.0005
29
+ lr_schedule: cosine
30
+ max_grad_norm: 10.0
31
+ min_lr_factor: 0.1
32
+ n_samples: 400000
33
+ saes:
34
+ dict_size_to_input_ratio: 60.0
35
+ k: null
36
+ pretrained_sae_paths: null
37
+ retrain_saes: false
38
+ sae_positions:
39
+ - blocks.2.hook_resid_pre
40
+ type_of_sparsifier: sae
41
+ save_dir: /mnt/ssd-interp/dan/sparsify/sparsify/scripts/train_tlens_saes/out
42
+ save_every_n_samples: null
43
+ seed: 0
44
+ tlens_model_name: gpt2-small
45
+ tlens_model_path: null
46
+ train_data:
47
+ column_name: input_ids
48
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
49
+ is_tokenized: true
50
+ n_ctx: 1024
51
+ seed: 0
52
+ split: train
53
+ streaming: true
54
+ tokenizer_name: gpt2
55
+ wandb_project: gpt2-layerwise_play
56
+ wandb_run_name: null
57
+ wandb_run_name_prefix: ''
58
+ warmup_samples: 20000
local_similar_ce_layer_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c49a66c6d9ae4f6ff636d887f31957caf1cffc6375078fcf9f7eb637a00b4ee4
3
+ size 283305112
local_similar_ce_layer_6.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 8
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 40000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl: null
22
+ out_to_in:
23
+ coeff: 1.0
24
+ out_to_orig: null
25
+ sparsity:
26
+ coeff: 4.0
27
+ p_norm: 1.0
28
+ lr: 0.0005
29
+ lr_schedule: cosine
30
+ max_grad_norm: 10.0
31
+ min_lr_factor: 0.1
32
+ n_samples: 400000
33
+ saes:
34
+ dict_size_to_input_ratio: 60.0
35
+ k: null
36
+ pretrained_sae_paths: null
37
+ retrain_saes: false
38
+ sae_positions:
39
+ - blocks.6.hook_resid_pre
40
+ type_of_sparsifier: sae
41
+ save_dir: /mnt/ssd-interp/dan/sparsify/sparsify/scripts/train_tlens_saes/out
42
+ save_every_n_samples: null
43
+ seed: 0
44
+ tlens_model_name: gpt2-small
45
+ tlens_model_path: null
46
+ train_data:
47
+ column_name: input_ids
48
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
49
+ is_tokenized: true
50
+ n_ctx: 1024
51
+ seed: 0
52
+ split: train
53
+ streaming: true
54
+ tokenizer_name: gpt2
55
+ wandb_project: gpt2-layerwise_play
56
+ wandb_run_name: null
57
+ wandb_run_name_prefix: ''
58
+ warmup_samples: 20000
local_similar_l0_layer_10.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a9f4178cf150c2402a9d71eae5322bc19d12af798d7f846971ed2ac6fc85506
3
+ size 283305120
local_similar_l0_layer_10.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 8
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 40000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl: null
22
+ out_to_in:
23
+ coeff: 1.0
24
+ out_to_orig: null
25
+ sparsity:
26
+ coeff: 10.0
27
+ p_norm: 1.0
28
+ lr: 0.0005
29
+ lr_schedule: cosine
30
+ max_grad_norm: 10.0
31
+ min_lr_factor: 0.1
32
+ n_samples: 400000
33
+ saes:
34
+ dict_size_to_input_ratio: 60.0
35
+ k: null
36
+ pretrained_sae_paths: null
37
+ retrain_saes: false
38
+ sae_positions:
39
+ - blocks.10.hook_resid_pre
40
+ type_of_sparsifier: sae
41
+ save_dir: /mnt/ssd-interp/dan/sparsify/sparsify/scripts/train_tlens_saes/out
42
+ save_every_n_samples: null
43
+ seed: 0
44
+ tlens_model_name: gpt2-small
45
+ tlens_model_path: null
46
+ train_data:
47
+ column_name: input_ids
48
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
49
+ is_tokenized: true
50
+ n_ctx: 1024
51
+ seed: 0
52
+ split: train
53
+ streaming: true
54
+ tokenizer_name: gpt2
55
+ wandb_project: gpt2-layerwise_play
56
+ wandb_run_name: null
57
+ wandb_run_name_prefix: ''
58
+ warmup_samples: 20000
local_similar_l0_layer_2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e24781476ae17f9c643177792075365b6f9c16963bc1e5d328730172e0394156
3
+ size 283305112
local_similar_l0_layer_2.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 8
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 40000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl: null
22
+ out_to_in:
23
+ coeff: 1.0
24
+ out_to_orig: null
25
+ sparsity:
26
+ coeff: 4.0
27
+ p_norm: 1.0
28
+ lr: 0.0005
29
+ lr_schedule: cosine
30
+ max_grad_norm: 10.0
31
+ min_lr_factor: 0.1
32
+ n_samples: 400000
33
+ saes:
34
+ dict_size_to_input_ratio: 60.0
35
+ k: null
36
+ pretrained_sae_paths: null
37
+ retrain_saes: false
38
+ sae_positions:
39
+ - blocks.2.hook_resid_pre
40
+ type_of_sparsifier: sae
41
+ save_dir: /mnt/ssd-interp/dan/sparsify/sparsify/scripts/train_tlens_saes/out
42
+ save_every_n_samples: null
43
+ seed: 0
44
+ tlens_model_name: gpt2-small
45
+ tlens_model_path: null
46
+ train_data:
47
+ column_name: input_ids
48
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
49
+ is_tokenized: true
50
+ n_ctx: 1024
51
+ seed: 0
52
+ split: train
53
+ streaming: true
54
+ tokenizer_name: gpt2
55
+ wandb_project: gpt2-layerwise_play
56
+ wandb_run_name: null
57
+ wandb_run_name_prefix: ''
58
+ warmup_samples: 20000
local_similar_l0_layer_6.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5517947946a84a2b26b787758da88897e15a6dd140c5a475e527da1aabe59dd7
3
+ size 283305112
local_similar_l0_layer_6.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ act_frequency_n_tokens: 500000
2
+ batch_size: 8
3
+ collect_act_frequency_every_n_samples: 40000
4
+ collect_output_metrics_every_n_samples: 0
5
+ cooldown_samples: 0
6
+ effective_batch_size: 16
7
+ eval_data:
8
+ column_name: input_ids
9
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
10
+ is_tokenized: true
11
+ n_ctx: 1024
12
+ seed: 0
13
+ split: train
14
+ streaming: true
15
+ tokenizer_name: gpt2
16
+ eval_every_n_samples: 40000
17
+ eval_n_samples: 500
18
+ log_every_n_grad_steps: 20
19
+ loss:
20
+ in_to_orig: null
21
+ logits_kl: null
22
+ out_to_in:
23
+ coeff: 1.0
24
+ out_to_orig: null
25
+ sparsity:
26
+ coeff: 6.0
27
+ p_norm: 1.0
28
+ lr: 0.0005
29
+ lr_schedule: cosine
30
+ max_grad_norm: 10.0
31
+ min_lr_factor: 0.1
32
+ n_samples: 400000
33
+ saes:
34
+ dict_size_to_input_ratio: 60.0
35
+ k: null
36
+ pretrained_sae_paths: null
37
+ retrain_saes: false
38
+ sae_positions:
39
+ - blocks.6.hook_resid_pre
40
+ type_of_sparsifier: sae
41
+ save_dir: /mnt/ssd-interp/dan/sparsify/sparsify/scripts/train_tlens_saes/out
42
+ save_every_n_samples: null
43
+ seed: 0
44
+ tlens_model_name: gpt2-small
45
+ tlens_model_path: null
46
+ train_data:
47
+ column_name: input_ids
48
+ dataset_name: apollo-research/Skylion007-openwebtext-tokenizer-gpt2
49
+ is_tokenized: true
50
+ n_ctx: 1024
51
+ seed: 0
52
+ split: train
53
+ streaming: true
54
+ tokenizer_name: gpt2
55
+ wandb_project: gpt2-layerwise_play
56
+ wandb_run_name: null
57
+ wandb_run_name_prefix: ''
58
+ warmup_samples: 20000