Add files using upload-large-folder tool
Browse files- .gitattributes +6 -0
- absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json +268 -0
- absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json +268 -0
- absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json +268 -0
- absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json +268 -0
- absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json +268 -0
- absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json +268 -0
- autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json +1 -0
- autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json +1 -0
- autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json +1 -0
- autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json +1 -0
- autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json +1 -0
- autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json +1 -0
- autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json +3 -0
- autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json +3 -0
- autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json +3 -0
- autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json +3 -0
- autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json +3 -0
- autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json +3 -0
- core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json +1 -0
- core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json +1 -0
- core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json +1 -0
- core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json +1 -0
- core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json +1 -0
- core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json +1 -0
- core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json +0 -0
- core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json +0 -0
- core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json +0 -0
- core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json +0 -0
- core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json +0 -0
- core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json +0 -0
- scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json +323 -0
- scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json +323 -0
- scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json +323 -0
- scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json +323 -0
- scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json +323 -0
- scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json +323 -0
- sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json +670 -0
- sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json +670 -0
- sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json +670 -0
- sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json +670 -0
- sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json +670 -0
- sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json +670 -0
- unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json +74 -0
- unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json +74 -0
- unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json +74 -0
- unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json +74 -0
- unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json +74 -0
- unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json +74 -0
.gitattributes
CHANGED
@@ -167,3 +167,9 @@ autointerp_with_generations/matryoshka_gemma-2-2b-16k-v2_MatryoshkaBatchTopKTrai
|
|
167 |
autointerp_with_generations/matryoshka_gemma-2-2b-16k-v2_MatryoshkaBatchTopKTrainer_65k_temp1000_google_gemma-2-2b_ctx1024_0117_resid_post_layer_12_trainer_0_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
168 |
autointerp_with_generations/matryoshka_gemma-2-2b-16k-v2_MatryoshkaBatchTopKTrainer_10_fixed_groups_google_gemma-2-2b_ctx1024_0117_resid_post_layer_12_trainer_1_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
169 |
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_batch_topk_65k_google_gemma-2-2b_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
167 |
autointerp_with_generations/matryoshka_gemma-2-2b-16k-v2_MatryoshkaBatchTopKTrainer_65k_temp1000_google_gemma-2-2b_ctx1024_0117_resid_post_layer_12_trainer_0_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
168 |
autointerp_with_generations/matryoshka_gemma-2-2b-16k-v2_MatryoshkaBatchTopKTrainer_10_fixed_groups_google_gemma-2-2b_ctx1024_0117_resid_post_layer_12_trainer_1_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
169 |
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_batch_topk_65k_google_gemma-2-2b_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
170 |
+
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
171 |
+
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
172 |
+
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
173 |
+
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
174 |
+
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
175 |
+
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json filter=lfs diff=lfs merge=lfs -text
|
absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,268 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "absorption_first_letter",
|
3 |
+
"eval_config": {
|
4 |
+
"model_name": "gemma-2-2b",
|
5 |
+
"random_seed": 42,
|
6 |
+
"f1_jump_threshold": 0.03,
|
7 |
+
"max_k_value": 10,
|
8 |
+
"prompt_template": "{word} has the first letter:",
|
9 |
+
"prompt_token_pos": -6,
|
10 |
+
"llm_batch_size": 32,
|
11 |
+
"llm_dtype": "bfloat16",
|
12 |
+
"k_sparse_probe_l1_decay": 0.01,
|
13 |
+
"k_sparse_probe_batch_size": 4096,
|
14 |
+
"k_sparse_probe_num_epochs": 50
|
15 |
+
},
|
16 |
+
"eval_id": "46f7f021-05a9-4488-a96f-7f354c70af6a",
|
17 |
+
"datetime_epoch_millis": 1737799424035,
|
18 |
+
"eval_result_metrics": {
|
19 |
+
"mean": {
|
20 |
+
"mean_absorption_fraction_score": 0.09458438446771604,
|
21 |
+
"mean_full_absorption_score": 0.13016930636592328,
|
22 |
+
"mean_num_split_features": 1.2692307692307692,
|
23 |
+
"std_dev_absorption_fraction_score": 0.07887489256096945,
|
24 |
+
"std_dev_full_absorption_score": 0.09727147892332459,
|
25 |
+
"std_dev_num_split_features": 0.827414880586145
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"eval_result_details": [
|
29 |
+
{
|
30 |
+
"first_letter": "a",
|
31 |
+
"mean_absorption_fraction": 0.1240437008257916,
|
32 |
+
"full_absorption_rate": 0.11189083820662768,
|
33 |
+
"num_full_absorption": 287,
|
34 |
+
"num_probe_true_positives": 2565,
|
35 |
+
"num_split_features": 1
|
36 |
+
},
|
37 |
+
{
|
38 |
+
"first_letter": "b",
|
39 |
+
"mean_absorption_fraction": 0.04824231927288525,
|
40 |
+
"full_absorption_rate": 0.07706535141800247,
|
41 |
+
"num_full_absorption": 125,
|
42 |
+
"num_probe_true_positives": 1622,
|
43 |
+
"num_split_features": 1
|
44 |
+
},
|
45 |
+
{
|
46 |
+
"first_letter": "c",
|
47 |
+
"mean_absorption_fraction": 0.19948124431575726,
|
48 |
+
"full_absorption_rate": 0.21801286633309508,
|
49 |
+
"num_full_absorption": 610,
|
50 |
+
"num_probe_true_positives": 2798,
|
51 |
+
"num_split_features": 4
|
52 |
+
},
|
53 |
+
{
|
54 |
+
"first_letter": "d",
|
55 |
+
"mean_absorption_fraction": 0.1466202617849163,
|
56 |
+
"full_absorption_rate": 0.2029673590504451,
|
57 |
+
"num_full_absorption": 342,
|
58 |
+
"num_probe_true_positives": 1685,
|
59 |
+
"num_split_features": 1
|
60 |
+
},
|
61 |
+
{
|
62 |
+
"first_letter": "e",
|
63 |
+
"mean_absorption_fraction": 0.29648549176335093,
|
64 |
+
"full_absorption_rate": 0.35918114143920593,
|
65 |
+
"num_full_absorption": 579,
|
66 |
+
"num_probe_true_positives": 1612,
|
67 |
+
"num_split_features": 4
|
68 |
+
},
|
69 |
+
{
|
70 |
+
"first_letter": "f",
|
71 |
+
"mean_absorption_fraction": 0.05002907238684505,
|
72 |
+
"full_absorption_rate": 0.060876623376623376,
|
73 |
+
"num_full_absorption": 75,
|
74 |
+
"num_probe_true_positives": 1232,
|
75 |
+
"num_split_features": 1
|
76 |
+
},
|
77 |
+
{
|
78 |
+
"first_letter": "g",
|
79 |
+
"mean_absorption_fraction": 0.14310827354955902,
|
80 |
+
"full_absorption_rate": 0.2041392285983067,
|
81 |
+
"num_full_absorption": 217,
|
82 |
+
"num_probe_true_positives": 1063,
|
83 |
+
"num_split_features": 1
|
84 |
+
},
|
85 |
+
{
|
86 |
+
"first_letter": "h",
|
87 |
+
"mean_absorption_fraction": 0.20666744991734257,
|
88 |
+
"full_absorption_rate": 0.2583170254403131,
|
89 |
+
"num_full_absorption": 264,
|
90 |
+
"num_probe_true_positives": 1022,
|
91 |
+
"num_split_features": 1
|
92 |
+
},
|
93 |
+
{
|
94 |
+
"first_letter": "i",
|
95 |
+
"mean_absorption_fraction": 0.02082145251831756,
|
96 |
+
"full_absorption_rate": 0.057125307125307126,
|
97 |
+
"num_full_absorption": 93,
|
98 |
+
"num_probe_true_positives": 1628,
|
99 |
+
"num_split_features": 1
|
100 |
+
},
|
101 |
+
{
|
102 |
+
"first_letter": "j",
|
103 |
+
"mean_absorption_fraction": 0.060406986257142754,
|
104 |
+
"full_absorption_rate": 0.1004566210045662,
|
105 |
+
"num_full_absorption": 44,
|
106 |
+
"num_probe_true_positives": 438,
|
107 |
+
"num_split_features": 1
|
108 |
+
},
|
109 |
+
{
|
110 |
+
"first_letter": "k",
|
111 |
+
"mean_absorption_fraction": 0.004578127898024072,
|
112 |
+
"full_absorption_rate": 0.027377521613832854,
|
113 |
+
"num_full_absorption": 19,
|
114 |
+
"num_probe_true_positives": 694,
|
115 |
+
"num_split_features": 1
|
116 |
+
},
|
117 |
+
{
|
118 |
+
"first_letter": "l",
|
119 |
+
"mean_absorption_fraction": 0.10657912825179995,
|
120 |
+
"full_absorption_rate": 0.17227564102564102,
|
121 |
+
"num_full_absorption": 215,
|
122 |
+
"num_probe_true_positives": 1248,
|
123 |
+
"num_split_features": 1
|
124 |
+
},
|
125 |
+
{
|
126 |
+
"first_letter": "m",
|
127 |
+
"mean_absorption_fraction": 0.25213309673289697,
|
128 |
+
"full_absorption_rate": 0.35584562996594776,
|
129 |
+
"num_full_absorption": 627,
|
130 |
+
"num_probe_true_positives": 1762,
|
131 |
+
"num_split_features": 1
|
132 |
+
},
|
133 |
+
{
|
134 |
+
"first_letter": "n",
|
135 |
+
"mean_absorption_fraction": 0.10849825960370131,
|
136 |
+
"full_absorption_rate": 0.13513513513513514,
|
137 |
+
"num_full_absorption": 115,
|
138 |
+
"num_probe_true_positives": 851,
|
139 |
+
"num_split_features": 1
|
140 |
+
},
|
141 |
+
{
|
142 |
+
"first_letter": "o",
|
143 |
+
"mean_absorption_fraction": 0.1319387636554577,
|
144 |
+
"full_absorption_rate": 0.2275390625,
|
145 |
+
"num_full_absorption": 233,
|
146 |
+
"num_probe_true_positives": 1024,
|
147 |
+
"num_split_features": 1
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"first_letter": "p",
|
151 |
+
"mean_absorption_fraction": 0.08315671739778407,
|
152 |
+
"full_absorption_rate": 0.1040587219343696,
|
153 |
+
"num_full_absorption": 241,
|
154 |
+
"num_probe_true_positives": 2316,
|
155 |
+
"num_split_features": 1
|
156 |
+
},
|
157 |
+
{
|
158 |
+
"first_letter": "q",
|
159 |
+
"mean_absorption_fraction": 0.0,
|
160 |
+
"full_absorption_rate": 0.005494505494505495,
|
161 |
+
"num_full_absorption": 1,
|
162 |
+
"num_probe_true_positives": 182,
|
163 |
+
"num_split_features": 1
|
164 |
+
},
|
165 |
+
{
|
166 |
+
"first_letter": "r",
|
167 |
+
"mean_absorption_fraction": 0.12568840190107153,
|
168 |
+
"full_absorption_rate": 0.15834348355663824,
|
169 |
+
"num_full_absorption": 260,
|
170 |
+
"num_probe_true_positives": 1642,
|
171 |
+
"num_split_features": 2
|
172 |
+
},
|
173 |
+
{
|
174 |
+
"first_letter": "s",
|
175 |
+
"mean_absorption_fraction": 0.09155612535225625,
|
176 |
+
"full_absorption_rate": 0.1314655172413793,
|
177 |
+
"num_full_absorption": 366,
|
178 |
+
"num_probe_true_positives": 2784,
|
179 |
+
"num_split_features": 1
|
180 |
+
},
|
181 |
+
{
|
182 |
+
"first_letter": "t",
|
183 |
+
"mean_absorption_fraction": 0.03506009203852256,
|
184 |
+
"full_absorption_rate": 0.05023364485981308,
|
185 |
+
"num_full_absorption": 86,
|
186 |
+
"num_probe_true_positives": 1712,
|
187 |
+
"num_split_features": 1
|
188 |
+
},
|
189 |
+
{
|
190 |
+
"first_letter": "u",
|
191 |
+
"mean_absorption_fraction": 0.10966650408949404,
|
192 |
+
"full_absorption_rate": 0.13850063532401524,
|
193 |
+
"num_full_absorption": 109,
|
194 |
+
"num_probe_true_positives": 787,
|
195 |
+
"num_split_features": 1
|
196 |
+
},
|
197 |
+
{
|
198 |
+
"first_letter": "v",
|
199 |
+
"mean_absorption_fraction": 0.015309366628156325,
|
200 |
+
"full_absorption_rate": 0.06355382619974059,
|
201 |
+
"num_full_absorption": 49,
|
202 |
+
"num_probe_true_positives": 771,
|
203 |
+
"num_split_features": 1
|
204 |
+
},
|
205 |
+
{
|
206 |
+
"first_letter": "w",
|
207 |
+
"mean_absorption_fraction": 0.03372208577383464,
|
208 |
+
"full_absorption_rate": 0.04481792717086835,
|
209 |
+
"num_full_absorption": 32,
|
210 |
+
"num_probe_true_positives": 714,
|
211 |
+
"num_split_features": 1
|
212 |
+
},
|
213 |
+
{
|
214 |
+
"first_letter": "x",
|
215 |
+
"mean_absorption_fraction": 0.02417116592097063,
|
216 |
+
"full_absorption_rate": 0.0,
|
217 |
+
"num_full_absorption": 0,
|
218 |
+
"num_probe_true_positives": 101,
|
219 |
+
"num_split_features": 1
|
220 |
+
},
|
221 |
+
{
|
222 |
+
"first_letter": "y",
|
223 |
+
"mean_absorption_fraction": 0.02955675657376613,
|
224 |
+
"full_absorption_rate": 0.049689440993788817,
|
225 |
+
"num_full_absorption": 8,
|
226 |
+
"num_probe_true_positives": 161,
|
227 |
+
"num_split_features": 1
|
228 |
+
},
|
229 |
+
{
|
230 |
+
"first_letter": "z",
|
231 |
+
"mean_absorption_fraction": 0.011673151750972763,
|
232 |
+
"full_absorption_rate": 0.07003891050583658,
|
233 |
+
"num_full_absorption": 18,
|
234 |
+
"num_probe_true_positives": 257,
|
235 |
+
"num_split_features": 1
|
236 |
+
}
|
237 |
+
],
|
238 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
239 |
+
"sae_lens_id": "custom_sae",
|
240 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0",
|
241 |
+
"sae_lens_version": "5.3.2",
|
242 |
+
"sae_cfg_dict": {
|
243 |
+
"model_name": "gemma-2-2b",
|
244 |
+
"d_in": 2304,
|
245 |
+
"d_sae": 16384,
|
246 |
+
"hook_layer": 12,
|
247 |
+
"hook_name": "blocks.12.hook_resid_post",
|
248 |
+
"context_size": null,
|
249 |
+
"hook_head_index": null,
|
250 |
+
"architecture": "matryoshka_batch_topk",
|
251 |
+
"apply_b_dec_to_input": null,
|
252 |
+
"finetuning_scaling_factor": null,
|
253 |
+
"activation_fn_str": "",
|
254 |
+
"prepend_bos": true,
|
255 |
+
"normalize_activations": "none",
|
256 |
+
"dtype": "bfloat16",
|
257 |
+
"device": "",
|
258 |
+
"dataset_path": "",
|
259 |
+
"dataset_trust_remote_code": true,
|
260 |
+
"seqpos_slice": [
|
261 |
+
null
|
262 |
+
],
|
263 |
+
"training_tokens": -100000,
|
264 |
+
"sae_lens_training_version": null,
|
265 |
+
"neuronpedia_id": null
|
266 |
+
},
|
267 |
+
"eval_result_unstructured": null
|
268 |
+
}
|
absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,268 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "absorption_first_letter",
|
3 |
+
"eval_config": {
|
4 |
+
"model_name": "gemma-2-2b",
|
5 |
+
"random_seed": 42,
|
6 |
+
"f1_jump_threshold": 0.03,
|
7 |
+
"max_k_value": 10,
|
8 |
+
"prompt_template": "{word} has the first letter:",
|
9 |
+
"prompt_token_pos": -6,
|
10 |
+
"llm_batch_size": 32,
|
11 |
+
"llm_dtype": "bfloat16",
|
12 |
+
"k_sparse_probe_l1_decay": 0.01,
|
13 |
+
"k_sparse_probe_batch_size": 4096,
|
14 |
+
"k_sparse_probe_num_epochs": 50
|
15 |
+
},
|
16 |
+
"eval_id": "4bd35244-b222-47cf-b1f9-211684736272",
|
17 |
+
"datetime_epoch_millis": 1737802155406,
|
18 |
+
"eval_result_metrics": {
|
19 |
+
"mean": {
|
20 |
+
"mean_absorption_fraction_score": 0.030515343203495613,
|
21 |
+
"mean_full_absorption_score": 0.035867854154501964,
|
22 |
+
"mean_num_split_features": 1.0384615384615385,
|
23 |
+
"std_dev_absorption_fraction_score": 0.05460457038794737,
|
24 |
+
"std_dev_full_absorption_score": 0.04302048742242669,
|
25 |
+
"std_dev_num_split_features": 0.19611613513818404
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"eval_result_details": [
|
29 |
+
{
|
30 |
+
"first_letter": "a",
|
31 |
+
"mean_absorption_fraction": 0.24795639270533665,
|
32 |
+
"full_absorption_rate": 0.15633528265107213,
|
33 |
+
"num_full_absorption": 401,
|
34 |
+
"num_probe_true_positives": 2565,
|
35 |
+
"num_split_features": 1
|
36 |
+
},
|
37 |
+
{
|
38 |
+
"first_letter": "b",
|
39 |
+
"mean_absorption_fraction": 0.0006165228113440197,
|
40 |
+
"full_absorption_rate": 0.006165228113440197,
|
41 |
+
"num_full_absorption": 10,
|
42 |
+
"num_probe_true_positives": 1622,
|
43 |
+
"num_split_features": 1
|
44 |
+
},
|
45 |
+
{
|
46 |
+
"first_letter": "c",
|
47 |
+
"mean_absorption_fraction": 0.0005247598087243948,
|
48 |
+
"full_absorption_rate": 0.004288777698355968,
|
49 |
+
"num_full_absorption": 12,
|
50 |
+
"num_probe_true_positives": 2798,
|
51 |
+
"num_split_features": 1
|
52 |
+
},
|
53 |
+
{
|
54 |
+
"first_letter": "d",
|
55 |
+
"mean_absorption_fraction": 0.008075285964749289,
|
56 |
+
"full_absorption_rate": 0.01483679525222552,
|
57 |
+
"num_full_absorption": 25,
|
58 |
+
"num_probe_true_positives": 1685,
|
59 |
+
"num_split_features": 1
|
60 |
+
},
|
61 |
+
{
|
62 |
+
"first_letter": "e",
|
63 |
+
"mean_absorption_fraction": 0.008615174704244508,
|
64 |
+
"full_absorption_rate": 0.026054590570719603,
|
65 |
+
"num_full_absorption": 42,
|
66 |
+
"num_probe_true_positives": 1612,
|
67 |
+
"num_split_features": 1
|
68 |
+
},
|
69 |
+
{
|
70 |
+
"first_letter": "f",
|
71 |
+
"mean_absorption_fraction": 0.06333141641449481,
|
72 |
+
"full_absorption_rate": 0.05925324675324675,
|
73 |
+
"num_full_absorption": 73,
|
74 |
+
"num_probe_true_positives": 1232,
|
75 |
+
"num_split_features": 1
|
76 |
+
},
|
77 |
+
{
|
78 |
+
"first_letter": "g",
|
79 |
+
"mean_absorption_fraction": 0.0378072186342135,
|
80 |
+
"full_absorption_rate": 0.043273753527751646,
|
81 |
+
"num_full_absorption": 46,
|
82 |
+
"num_probe_true_positives": 1063,
|
83 |
+
"num_split_features": 1
|
84 |
+
},
|
85 |
+
{
|
86 |
+
"first_letter": "h",
|
87 |
+
"mean_absorption_fraction": 0.03992355071730363,
|
88 |
+
"full_absorption_rate": 0.04794520547945205,
|
89 |
+
"num_full_absorption": 49,
|
90 |
+
"num_probe_true_positives": 1022,
|
91 |
+
"num_split_features": 1
|
92 |
+
},
|
93 |
+
{
|
94 |
+
"first_letter": "i",
|
95 |
+
"mean_absorption_fraction": 0.014734260170669538,
|
96 |
+
"full_absorption_rate": 0.04176904176904177,
|
97 |
+
"num_full_absorption": 68,
|
98 |
+
"num_probe_true_positives": 1628,
|
99 |
+
"num_split_features": 1
|
100 |
+
},
|
101 |
+
{
|
102 |
+
"first_letter": "j",
|
103 |
+
"mean_absorption_fraction": 0.0030458098177117703,
|
104 |
+
"full_absorption_rate": 0.0136986301369863,
|
105 |
+
"num_full_absorption": 6,
|
106 |
+
"num_probe_true_positives": 438,
|
107 |
+
"num_split_features": 1
|
108 |
+
},
|
109 |
+
{
|
110 |
+
"first_letter": "k",
|
111 |
+
"mean_absorption_fraction": 0.04603562428706451,
|
112 |
+
"full_absorption_rate": 0.06051873198847262,
|
113 |
+
"num_full_absorption": 42,
|
114 |
+
"num_probe_true_positives": 694,
|
115 |
+
"num_split_features": 1
|
116 |
+
},
|
117 |
+
{
|
118 |
+
"first_letter": "l",
|
119 |
+
"mean_absorption_fraction": 0.0,
|
120 |
+
"full_absorption_rate": 0.008012820512820512,
|
121 |
+
"num_full_absorption": 10,
|
122 |
+
"num_probe_true_positives": 1248,
|
123 |
+
"num_split_features": 1
|
124 |
+
},
|
125 |
+
{
|
126 |
+
"first_letter": "m",
|
127 |
+
"mean_absorption_fraction": 0.0,
|
128 |
+
"full_absorption_rate": 0.004540295119182747,
|
129 |
+
"num_full_absorption": 8,
|
130 |
+
"num_probe_true_positives": 1762,
|
131 |
+
"num_split_features": 1
|
132 |
+
},
|
133 |
+
{
|
134 |
+
"first_letter": "n",
|
135 |
+
"mean_absorption_fraction": 0.014503582233943687,
|
136 |
+
"full_absorption_rate": 0.01880141010575793,
|
137 |
+
"num_full_absorption": 16,
|
138 |
+
"num_probe_true_positives": 851,
|
139 |
+
"num_split_features": 1
|
140 |
+
},
|
141 |
+
{
|
142 |
+
"first_letter": "o",
|
143 |
+
"mean_absorption_fraction": 0.008062183272034208,
|
144 |
+
"full_absorption_rate": 0.0615234375,
|
145 |
+
"num_full_absorption": 63,
|
146 |
+
"num_probe_true_positives": 1024,
|
147 |
+
"num_split_features": 1
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"first_letter": "p",
|
151 |
+
"mean_absorption_fraction": 0.007621033040612787,
|
152 |
+
"full_absorption_rate": 0.015112262521588947,
|
153 |
+
"num_full_absorption": 35,
|
154 |
+
"num_probe_true_positives": 2316,
|
155 |
+
"num_split_features": 1
|
156 |
+
},
|
157 |
+
{
|
158 |
+
"first_letter": "q",
|
159 |
+
"mean_absorption_fraction": 0.0,
|
160 |
+
"full_absorption_rate": 0.005494505494505495,
|
161 |
+
"num_full_absorption": 1,
|
162 |
+
"num_probe_true_positives": 182,
|
163 |
+
"num_split_features": 1
|
164 |
+
},
|
165 |
+
{
|
166 |
+
"first_letter": "r",
|
167 |
+
"mean_absorption_fraction": 0.0007007994110090998,
|
168 |
+
"full_absorption_rate": 0.0030450669914738123,
|
169 |
+
"num_full_absorption": 5,
|
170 |
+
"num_probe_true_positives": 1642,
|
171 |
+
"num_split_features": 1
|
172 |
+
},
|
173 |
+
{
|
174 |
+
"first_letter": "s",
|
175 |
+
"mean_absorption_fraction": 0.13911983905775224,
|
176 |
+
"full_absorption_rate": 0.15768678160919541,
|
177 |
+
"num_full_absorption": 439,
|
178 |
+
"num_probe_true_positives": 2784,
|
179 |
+
"num_split_features": 1
|
180 |
+
},
|
181 |
+
{
|
182 |
+
"first_letter": "t",
|
183 |
+
"mean_absorption_fraction": 0.010533912113956346,
|
184 |
+
"full_absorption_rate": 0.011098130841121495,
|
185 |
+
"num_full_absorption": 19,
|
186 |
+
"num_probe_true_positives": 1712,
|
187 |
+
"num_split_features": 1
|
188 |
+
},
|
189 |
+
{
|
190 |
+
"first_letter": "u",
|
191 |
+
"mean_absorption_fraction": 0.003178350186885033,
|
192 |
+
"full_absorption_rate": 0.010165184243964422,
|
193 |
+
"num_full_absorption": 8,
|
194 |
+
"num_probe_true_positives": 787,
|
195 |
+
"num_split_features": 2
|
196 |
+
},
|
197 |
+
{
|
198 |
+
"first_letter": "v",
|
199 |
+
"mean_absorption_fraction": 0.0,
|
200 |
+
"full_absorption_rate": 0.005188067444876783,
|
201 |
+
"num_full_absorption": 4,
|
202 |
+
"num_probe_true_positives": 771,
|
203 |
+
"num_split_features": 1
|
204 |
+
},
|
205 |
+
{
|
206 |
+
"first_letter": "w",
|
207 |
+
"mean_absorption_fraction": 0.009737462643756158,
|
208 |
+
"full_absorption_rate": 0.014005602240896359,
|
209 |
+
"num_full_absorption": 10,
|
210 |
+
"num_probe_true_positives": 714,
|
211 |
+
"num_split_features": 1
|
212 |
+
},
|
213 |
+
{
|
214 |
+
"first_letter": "x",
|
215 |
+
"mean_absorption_fraction": 0.04181263509225596,
|
216 |
+
"full_absorption_rate": 0.0,
|
217 |
+
"num_full_absorption": 0,
|
218 |
+
"num_probe_true_positives": 101,
|
219 |
+
"num_split_features": 1
|
220 |
+
},
|
221 |
+
{
|
222 |
+
"first_letter": "y",
|
223 |
+
"mean_absorption_fraction": 0.07744156226402121,
|
224 |
+
"full_absorption_rate": 0.09316770186335403,
|
225 |
+
"num_full_absorption": 15,
|
226 |
+
"num_probe_true_positives": 161,
|
227 |
+
"num_split_features": 1
|
228 |
+
},
|
229 |
+
{
|
230 |
+
"first_letter": "z",
|
231 |
+
"mean_absorption_fraction": 0.010021547938802592,
|
232 |
+
"full_absorption_rate": 0.05058365758754864,
|
233 |
+
"num_full_absorption": 13,
|
234 |
+
"num_probe_true_positives": 257,
|
235 |
+
"num_split_features": 1
|
236 |
+
}
|
237 |
+
],
|
238 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
239 |
+
"sae_lens_id": "custom_sae",
|
240 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1",
|
241 |
+
"sae_lens_version": "5.3.2",
|
242 |
+
"sae_cfg_dict": {
|
243 |
+
"model_name": "gemma-2-2b",
|
244 |
+
"d_in": 2304,
|
245 |
+
"d_sae": 16384,
|
246 |
+
"hook_layer": 12,
|
247 |
+
"hook_name": "blocks.12.hook_resid_post",
|
248 |
+
"context_size": null,
|
249 |
+
"hook_head_index": null,
|
250 |
+
"architecture": "matryoshka_batch_topk",
|
251 |
+
"apply_b_dec_to_input": null,
|
252 |
+
"finetuning_scaling_factor": null,
|
253 |
+
"activation_fn_str": "",
|
254 |
+
"prepend_bos": true,
|
255 |
+
"normalize_activations": "none",
|
256 |
+
"dtype": "bfloat16",
|
257 |
+
"device": "",
|
258 |
+
"dataset_path": "",
|
259 |
+
"dataset_trust_remote_code": true,
|
260 |
+
"seqpos_slice": [
|
261 |
+
null
|
262 |
+
],
|
263 |
+
"training_tokens": -100000,
|
264 |
+
"sae_lens_training_version": null,
|
265 |
+
"neuronpedia_id": null
|
266 |
+
},
|
267 |
+
"eval_result_unstructured": null
|
268 |
+
}
|
absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,268 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "absorption_first_letter",
|
3 |
+
"eval_config": {
|
4 |
+
"model_name": "gemma-2-2b",
|
5 |
+
"random_seed": 42,
|
6 |
+
"f1_jump_threshold": 0.03,
|
7 |
+
"max_k_value": 10,
|
8 |
+
"prompt_template": "{word} has the first letter:",
|
9 |
+
"prompt_token_pos": -6,
|
10 |
+
"llm_batch_size": 32,
|
11 |
+
"llm_dtype": "bfloat16",
|
12 |
+
"k_sparse_probe_l1_decay": 0.01,
|
13 |
+
"k_sparse_probe_batch_size": 4096,
|
14 |
+
"k_sparse_probe_num_epochs": 50
|
15 |
+
},
|
16 |
+
"eval_id": "9cbe10f7-5e21-4141-a697-d87174bf326c",
|
17 |
+
"datetime_epoch_millis": 1737804972054,
|
18 |
+
"eval_result_metrics": {
|
19 |
+
"mean": {
|
20 |
+
"mean_absorption_fraction_score": 0.03106953109960984,
|
21 |
+
"mean_full_absorption_score": 0.012854656905879383,
|
22 |
+
"mean_num_split_features": 1.0769230769230769,
|
23 |
+
"std_dev_absorption_fraction_score": 0.07575219388605692,
|
24 |
+
"std_dev_full_absorption_score": 0.024754535069899743,
|
25 |
+
"std_dev_num_split_features": 0.271746488194703
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"eval_result_details": [
|
29 |
+
{
|
30 |
+
"first_letter": "a",
|
31 |
+
"mean_absorption_fraction": 0.0065242535617624445,
|
32 |
+
"full_absorption_rate": 0.001949317738791423,
|
33 |
+
"num_full_absorption": 5,
|
34 |
+
"num_probe_true_positives": 2565,
|
35 |
+
"num_split_features": 1
|
36 |
+
},
|
37 |
+
{
|
38 |
+
"first_letter": "b",
|
39 |
+
"mean_absorption_fraction": 0.0,
|
40 |
+
"full_absorption_rate": 0.0006165228113440197,
|
41 |
+
"num_full_absorption": 1,
|
42 |
+
"num_probe_true_positives": 1622,
|
43 |
+
"num_split_features": 1
|
44 |
+
},
|
45 |
+
{
|
46 |
+
"first_letter": "c",
|
47 |
+
"mean_absorption_fraction": 0.0,
|
48 |
+
"full_absorption_rate": 0.0007147962830593281,
|
49 |
+
"num_full_absorption": 2,
|
50 |
+
"num_probe_true_positives": 2798,
|
51 |
+
"num_split_features": 1
|
52 |
+
},
|
53 |
+
{
|
54 |
+
"first_letter": "d",
|
55 |
+
"mean_absorption_fraction": 0.0074862477164824885,
|
56 |
+
"full_absorption_rate": 0.002967359050445104,
|
57 |
+
"num_full_absorption": 5,
|
58 |
+
"num_probe_true_positives": 1685,
|
59 |
+
"num_split_features": 1
|
60 |
+
},
|
61 |
+
{
|
62 |
+
"first_letter": "e",
|
63 |
+
"mean_absorption_fraction": 0.12027184537311117,
|
64 |
+
"full_absorption_rate": 0.06265508684863523,
|
65 |
+
"num_full_absorption": 101,
|
66 |
+
"num_probe_true_positives": 1612,
|
67 |
+
"num_split_features": 1
|
68 |
+
},
|
69 |
+
{
|
70 |
+
"first_letter": "f",
|
71 |
+
"mean_absorption_fraction": 8.533749122617099e-05,
|
72 |
+
"full_absorption_rate": 0.0,
|
73 |
+
"num_full_absorption": 0,
|
74 |
+
"num_probe_true_positives": 1232,
|
75 |
+
"num_split_features": 1
|
76 |
+
},
|
77 |
+
{
|
78 |
+
"first_letter": "g",
|
79 |
+
"mean_absorption_fraction": 0.003381354643834261,
|
80 |
+
"full_absorption_rate": 0.004703668861712135,
|
81 |
+
"num_full_absorption": 5,
|
82 |
+
"num_probe_true_positives": 1063,
|
83 |
+
"num_split_features": 2
|
84 |
+
},
|
85 |
+
{
|
86 |
+
"first_letter": "h",
|
87 |
+
"mean_absorption_fraction": 0.008193616707506776,
|
88 |
+
"full_absorption_rate": 0.003913894324853229,
|
89 |
+
"num_full_absorption": 4,
|
90 |
+
"num_probe_true_positives": 1022,
|
91 |
+
"num_split_features": 1
|
92 |
+
},
|
93 |
+
{
|
94 |
+
"first_letter": "i",
|
95 |
+
"mean_absorption_fraction": 0.10861916195323115,
|
96 |
+
"full_absorption_rate": 0.07309582309582309,
|
97 |
+
"num_full_absorption": 119,
|
98 |
+
"num_probe_true_positives": 1628,
|
99 |
+
"num_split_features": 1
|
100 |
+
},
|
101 |
+
{
|
102 |
+
"first_letter": "j",
|
103 |
+
"mean_absorption_fraction": 0.004907091505762464,
|
104 |
+
"full_absorption_rate": 0.0091324200913242,
|
105 |
+
"num_full_absorption": 4,
|
106 |
+
"num_probe_true_positives": 438,
|
107 |
+
"num_split_features": 1
|
108 |
+
},
|
109 |
+
{
|
110 |
+
"first_letter": "k",
|
111 |
+
"mean_absorption_fraction": 0.3657601721329182,
|
112 |
+
"full_absorption_rate": 0.09654178674351585,
|
113 |
+
"num_full_absorption": 67,
|
114 |
+
"num_probe_true_positives": 694,
|
115 |
+
"num_split_features": 1
|
116 |
+
},
|
117 |
+
{
|
118 |
+
"first_letter": "l",
|
119 |
+
"mean_absorption_fraction": 0.0,
|
120 |
+
"full_absorption_rate": 0.0008012820512820513,
|
121 |
+
"num_full_absorption": 1,
|
122 |
+
"num_probe_true_positives": 1248,
|
123 |
+
"num_split_features": 1
|
124 |
+
},
|
125 |
+
{
|
126 |
+
"first_letter": "m",
|
127 |
+
"mean_absorption_fraction": 0.0,
|
128 |
+
"full_absorption_rate": 0.004540295119182747,
|
129 |
+
"num_full_absorption": 8,
|
130 |
+
"num_probe_true_positives": 1762,
|
131 |
+
"num_split_features": 1
|
132 |
+
},
|
133 |
+
{
|
134 |
+
"first_letter": "n",
|
135 |
+
"mean_absorption_fraction": 0.016536295172747958,
|
136 |
+
"full_absorption_rate": 0.01762632197414806,
|
137 |
+
"num_full_absorption": 15,
|
138 |
+
"num_probe_true_positives": 851,
|
139 |
+
"num_split_features": 1
|
140 |
+
},
|
141 |
+
{
|
142 |
+
"first_letter": "o",
|
143 |
+
"mean_absorption_fraction": 0.00484391170242484,
|
144 |
+
"full_absorption_rate": 0.00390625,
|
145 |
+
"num_full_absorption": 4,
|
146 |
+
"num_probe_true_positives": 1024,
|
147 |
+
"num_split_features": 1
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"first_letter": "p",
|
151 |
+
"mean_absorption_fraction": 0.0018436539428919063,
|
152 |
+
"full_absorption_rate": 0.0017271157167530224,
|
153 |
+
"num_full_absorption": 4,
|
154 |
+
"num_probe_true_positives": 2316,
|
155 |
+
"num_split_features": 1
|
156 |
+
},
|
157 |
+
{
|
158 |
+
"first_letter": "q",
|
159 |
+
"mean_absorption_fraction": 0.0008416303249357576,
|
160 |
+
"full_absorption_rate": 0.005494505494505495,
|
161 |
+
"num_full_absorption": 1,
|
162 |
+
"num_probe_true_positives": 182,
|
163 |
+
"num_split_features": 1
|
164 |
+
},
|
165 |
+
{
|
166 |
+
"first_letter": "r",
|
167 |
+
"mean_absorption_fraction": 7.436500627589195e-05,
|
168 |
+
"full_absorption_rate": 0.001218026796589525,
|
169 |
+
"num_full_absorption": 2,
|
170 |
+
"num_probe_true_positives": 1642,
|
171 |
+
"num_split_features": 1
|
172 |
+
},
|
173 |
+
{
|
174 |
+
"first_letter": "s",
|
175 |
+
"mean_absorption_fraction": 0.0036346646760435732,
|
176 |
+
"full_absorption_rate": 0.0028735632183908046,
|
177 |
+
"num_full_absorption": 8,
|
178 |
+
"num_probe_true_positives": 2784,
|
179 |
+
"num_split_features": 1
|
180 |
+
},
|
181 |
+
{
|
182 |
+
"first_letter": "t",
|
183 |
+
"mean_absorption_fraction": 0.009130431247648941,
|
184 |
+
"full_absorption_rate": 0.007593457943925234,
|
185 |
+
"num_full_absorption": 13,
|
186 |
+
"num_probe_true_positives": 1712,
|
187 |
+
"num_split_features": 1
|
188 |
+
},
|
189 |
+
{
|
190 |
+
"first_letter": "u",
|
191 |
+
"mean_absorption_fraction": 0.0449379501080026,
|
192 |
+
"full_absorption_rate": 0.005082592121982211,
|
193 |
+
"num_full_absorption": 4,
|
194 |
+
"num_probe_true_positives": 787,
|
195 |
+
"num_split_features": 1
|
196 |
+
},
|
197 |
+
{
|
198 |
+
"first_letter": "v",
|
199 |
+
"mean_absorption_fraction": 0.0,
|
200 |
+
"full_absorption_rate": 0.0,
|
201 |
+
"num_full_absorption": 0,
|
202 |
+
"num_probe_true_positives": 771,
|
203 |
+
"num_split_features": 1
|
204 |
+
},
|
205 |
+
{
|
206 |
+
"first_letter": "w",
|
207 |
+
"mean_absorption_fraction": 0.0052064341754823485,
|
208 |
+
"full_absorption_rate": 0.0014005602240896359,
|
209 |
+
"num_full_absorption": 1,
|
210 |
+
"num_probe_true_positives": 714,
|
211 |
+
"num_split_features": 1
|
212 |
+
},
|
213 |
+
{
|
214 |
+
"first_letter": "x",
|
215 |
+
"mean_absorption_fraction": 0.07513900163748029,
|
216 |
+
"full_absorption_rate": 0.0,
|
217 |
+
"num_full_absorption": 0,
|
218 |
+
"num_probe_true_positives": 101,
|
219 |
+
"num_split_features": 1
|
220 |
+
},
|
221 |
+
{
|
222 |
+
"first_letter": "y",
|
223 |
+
"mean_absorption_fraction": 0.01293314128922243,
|
224 |
+
"full_absorption_rate": 0.006211180124223602,
|
225 |
+
"num_full_absorption": 1,
|
226 |
+
"num_probe_true_positives": 161,
|
227 |
+
"num_split_features": 2
|
228 |
+
},
|
229 |
+
{
|
230 |
+
"first_letter": "z",
|
231 |
+
"mean_absorption_fraction": 0.007457248220864149,
|
232 |
+
"full_absorption_rate": 0.019455252918287938,
|
233 |
+
"num_full_absorption": 5,
|
234 |
+
"num_probe_true_positives": 257,
|
235 |
+
"num_split_features": 1
|
236 |
+
}
|
237 |
+
],
|
238 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
239 |
+
"sae_lens_id": "custom_sae",
|
240 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2",
|
241 |
+
"sae_lens_version": "5.3.2",
|
242 |
+
"sae_cfg_dict": {
|
243 |
+
"model_name": "gemma-2-2b",
|
244 |
+
"d_in": 2304,
|
245 |
+
"d_sae": 16384,
|
246 |
+
"hook_layer": 12,
|
247 |
+
"hook_name": "blocks.12.hook_resid_post",
|
248 |
+
"context_size": null,
|
249 |
+
"hook_head_index": null,
|
250 |
+
"architecture": "matryoshka_batch_topk",
|
251 |
+
"apply_b_dec_to_input": null,
|
252 |
+
"finetuning_scaling_factor": null,
|
253 |
+
"activation_fn_str": "",
|
254 |
+
"prepend_bos": true,
|
255 |
+
"normalize_activations": "none",
|
256 |
+
"dtype": "bfloat16",
|
257 |
+
"device": "",
|
258 |
+
"dataset_path": "",
|
259 |
+
"dataset_trust_remote_code": true,
|
260 |
+
"seqpos_slice": [
|
261 |
+
null
|
262 |
+
],
|
263 |
+
"training_tokens": -100000,
|
264 |
+
"sae_lens_training_version": null,
|
265 |
+
"neuronpedia_id": null
|
266 |
+
},
|
267 |
+
"eval_result_unstructured": null
|
268 |
+
}
|
absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,268 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "absorption_first_letter",
|
3 |
+
"eval_config": {
|
4 |
+
"model_name": "gemma-2-2b",
|
5 |
+
"random_seed": 42,
|
6 |
+
"f1_jump_threshold": 0.03,
|
7 |
+
"max_k_value": 10,
|
8 |
+
"prompt_template": "{word} has the first letter:",
|
9 |
+
"prompt_token_pos": -6,
|
10 |
+
"llm_batch_size": 32,
|
11 |
+
"llm_dtype": "bfloat16",
|
12 |
+
"k_sparse_probe_l1_decay": 0.01,
|
13 |
+
"k_sparse_probe_batch_size": 4096,
|
14 |
+
"k_sparse_probe_num_epochs": 50
|
15 |
+
},
|
16 |
+
"eval_id": "7c92dbf9-3524-4134-ae46-8ade1007c5e6",
|
17 |
+
"datetime_epoch_millis": 1737807807590,
|
18 |
+
"eval_result_metrics": {
|
19 |
+
"mean": {
|
20 |
+
"mean_absorption_fraction_score": 0.031425759865975664,
|
21 |
+
"mean_full_absorption_score": 0.00798730220887853,
|
22 |
+
"mean_num_split_features": 1.2692307692307692,
|
23 |
+
"std_dev_absorption_fraction_score": 0.08877416427897664,
|
24 |
+
"std_dev_full_absorption_score": 0.02236479529501252,
|
25 |
+
"std_dev_num_split_features": 0.533493565673837
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"eval_result_details": [
|
29 |
+
{
|
30 |
+
"first_letter": "a",
|
31 |
+
"mean_absorption_fraction": 0.002968917098976069,
|
32 |
+
"full_absorption_rate": 0.0,
|
33 |
+
"num_full_absorption": 0,
|
34 |
+
"num_probe_true_positives": 2565,
|
35 |
+
"num_split_features": 2
|
36 |
+
},
|
37 |
+
{
|
38 |
+
"first_letter": "b",
|
39 |
+
"mean_absorption_fraction": 8.305828714954895e-05,
|
40 |
+
"full_absorption_rate": 0.0006165228113440197,
|
41 |
+
"num_full_absorption": 1,
|
42 |
+
"num_probe_true_positives": 1622,
|
43 |
+
"num_split_features": 1
|
44 |
+
},
|
45 |
+
{
|
46 |
+
"first_letter": "c",
|
47 |
+
"mean_absorption_fraction": 0.0,
|
48 |
+
"full_absorption_rate": 0.0,
|
49 |
+
"num_full_absorption": 0,
|
50 |
+
"num_probe_true_positives": 2798,
|
51 |
+
"num_split_features": 1
|
52 |
+
},
|
53 |
+
{
|
54 |
+
"first_letter": "d",
|
55 |
+
"mean_absorption_fraction": 0.0038610564653031487,
|
56 |
+
"full_absorption_rate": 0.002967359050445104,
|
57 |
+
"num_full_absorption": 5,
|
58 |
+
"num_probe_true_positives": 1685,
|
59 |
+
"num_split_features": 1
|
60 |
+
},
|
61 |
+
{
|
62 |
+
"first_letter": "e",
|
63 |
+
"mean_absorption_fraction": 0.00020673381024651776,
|
64 |
+
"full_absorption_rate": 0.0,
|
65 |
+
"num_full_absorption": 0,
|
66 |
+
"num_probe_true_positives": 1612,
|
67 |
+
"num_split_features": 3
|
68 |
+
},
|
69 |
+
{
|
70 |
+
"first_letter": "f",
|
71 |
+
"mean_absorption_fraction": 0.0043141083006187535,
|
72 |
+
"full_absorption_rate": 0.0,
|
73 |
+
"num_full_absorption": 0,
|
74 |
+
"num_probe_true_positives": 1232,
|
75 |
+
"num_split_features": 1
|
76 |
+
},
|
77 |
+
{
|
78 |
+
"first_letter": "g",
|
79 |
+
"mean_absorption_fraction": 0.0010555334601028094,
|
80 |
+
"full_absorption_rate": 0.0018814675446848542,
|
81 |
+
"num_full_absorption": 2,
|
82 |
+
"num_probe_true_positives": 1063,
|
83 |
+
"num_split_features": 2
|
84 |
+
},
|
85 |
+
{
|
86 |
+
"first_letter": "h",
|
87 |
+
"mean_absorption_fraction": 0.0012612387639860632,
|
88 |
+
"full_absorption_rate": 0.0009784735812133072,
|
89 |
+
"num_full_absorption": 1,
|
90 |
+
"num_probe_true_positives": 1022,
|
91 |
+
"num_split_features": 1
|
92 |
+
},
|
93 |
+
{
|
94 |
+
"first_letter": "i",
|
95 |
+
"mean_absorption_fraction": 0.0008096975471463345,
|
96 |
+
"full_absorption_rate": 0.0,
|
97 |
+
"num_full_absorption": 0,
|
98 |
+
"num_probe_true_positives": 1628,
|
99 |
+
"num_split_features": 2
|
100 |
+
},
|
101 |
+
{
|
102 |
+
"first_letter": "j",
|
103 |
+
"mean_absorption_fraction": 0.0010560518434486158,
|
104 |
+
"full_absorption_rate": 0.0,
|
105 |
+
"num_full_absorption": 0,
|
106 |
+
"num_probe_true_positives": 438,
|
107 |
+
"num_split_features": 2
|
108 |
+
},
|
109 |
+
{
|
110 |
+
"first_letter": "k",
|
111 |
+
"mean_absorption_fraction": 0.39448351831277473,
|
112 |
+
"full_absorption_rate": 0.0936599423631124,
|
113 |
+
"num_full_absorption": 65,
|
114 |
+
"num_probe_true_positives": 694,
|
115 |
+
"num_split_features": 1
|
116 |
+
},
|
117 |
+
{
|
118 |
+
"first_letter": "l",
|
119 |
+
"mean_absorption_fraction": 0.0011242061308802099,
|
120 |
+
"full_absorption_rate": 0.008814102564102564,
|
121 |
+
"num_full_absorption": 11,
|
122 |
+
"num_probe_true_positives": 1248,
|
123 |
+
"num_split_features": 1
|
124 |
+
},
|
125 |
+
{
|
126 |
+
"first_letter": "m",
|
127 |
+
"mean_absorption_fraction": 0.0,
|
128 |
+
"full_absorption_rate": 0.006242905788876277,
|
129 |
+
"num_full_absorption": 11,
|
130 |
+
"num_probe_true_positives": 1762,
|
131 |
+
"num_split_features": 1
|
132 |
+
},
|
133 |
+
{
|
134 |
+
"first_letter": "n",
|
135 |
+
"mean_absorption_fraction": 0.0,
|
136 |
+
"full_absorption_rate": 0.0,
|
137 |
+
"num_full_absorption": 0,
|
138 |
+
"num_probe_true_positives": 851,
|
139 |
+
"num_split_features": 1
|
140 |
+
},
|
141 |
+
{
|
142 |
+
"first_letter": "o",
|
143 |
+
"mean_absorption_fraction": 0.0,
|
144 |
+
"full_absorption_rate": 0.0,
|
145 |
+
"num_full_absorption": 0,
|
146 |
+
"num_probe_true_positives": 1024,
|
147 |
+
"num_split_features": 2
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"first_letter": "p",
|
151 |
+
"mean_absorption_fraction": 0.00010267550041210845,
|
152 |
+
"full_absorption_rate": 0.0012953367875647669,
|
153 |
+
"num_full_absorption": 3,
|
154 |
+
"num_probe_true_positives": 2316,
|
155 |
+
"num_split_features": 1
|
156 |
+
},
|
157 |
+
{
|
158 |
+
"first_letter": "q",
|
159 |
+
"mean_absorption_fraction": 0.00308986483396796,
|
160 |
+
"full_absorption_rate": 0.0,
|
161 |
+
"num_full_absorption": 0,
|
162 |
+
"num_probe_true_positives": 182,
|
163 |
+
"num_split_features": 1
|
164 |
+
},
|
165 |
+
{
|
166 |
+
"first_letter": "r",
|
167 |
+
"mean_absorption_fraction": 0.0,
|
168 |
+
"full_absorption_rate": 0.0,
|
169 |
+
"num_full_absorption": 0,
|
170 |
+
"num_probe_true_positives": 1642,
|
171 |
+
"num_split_features": 1
|
172 |
+
},
|
173 |
+
{
|
174 |
+
"first_letter": "s",
|
175 |
+
"mean_absorption_fraction": 5.2977770731292506e-05,
|
176 |
+
"full_absorption_rate": 0.0,
|
177 |
+
"num_full_absorption": 0,
|
178 |
+
"num_probe_true_positives": 2784,
|
179 |
+
"num_split_features": 1
|
180 |
+
},
|
181 |
+
{
|
182 |
+
"first_letter": "t",
|
183 |
+
"mean_absorption_fraction": 0.00010220640700548763,
|
184 |
+
"full_absorption_rate": 0.0,
|
185 |
+
"num_full_absorption": 0,
|
186 |
+
"num_probe_true_positives": 1712,
|
187 |
+
"num_split_features": 1
|
188 |
+
},
|
189 |
+
{
|
190 |
+
"first_letter": "u",
|
191 |
+
"mean_absorption_fraction": 0.024324725109985613,
|
192 |
+
"full_absorption_rate": 0.0025412960609911056,
|
193 |
+
"num_full_absorption": 2,
|
194 |
+
"num_probe_true_positives": 787,
|
195 |
+
"num_split_features": 1
|
196 |
+
},
|
197 |
+
{
|
198 |
+
"first_letter": "v",
|
199 |
+
"mean_absorption_fraction": 0.0034945719681779308,
|
200 |
+
"full_absorption_rate": 0.0,
|
201 |
+
"num_full_absorption": 0,
|
202 |
+
"num_probe_true_positives": 771,
|
203 |
+
"num_split_features": 1
|
204 |
+
},
|
205 |
+
{
|
206 |
+
"first_letter": "w",
|
207 |
+
"mean_absorption_fraction": 0.00400852548032978,
|
208 |
+
"full_absorption_rate": 0.0,
|
209 |
+
"num_full_absorption": 0,
|
210 |
+
"num_probe_true_positives": 714,
|
211 |
+
"num_split_features": 1
|
212 |
+
},
|
213 |
+
{
|
214 |
+
"first_letter": "x",
|
215 |
+
"mean_absorption_fraction": 0.06699006775636132,
|
216 |
+
"full_absorption_rate": 0.0,
|
217 |
+
"num_full_absorption": 0,
|
218 |
+
"num_probe_true_positives": 101,
|
219 |
+
"num_split_features": 1
|
220 |
+
},
|
221 |
+
{
|
222 |
+
"first_letter": "y",
|
223 |
+
"mean_absorption_fraction": 0.06294404836876093,
|
224 |
+
"full_absorption_rate": 0.018633540372670808,
|
225 |
+
"num_full_absorption": 3,
|
226 |
+
"num_probe_true_positives": 161,
|
227 |
+
"num_split_features": 1
|
228 |
+
},
|
229 |
+
{
|
230 |
+
"first_letter": "z",
|
231 |
+
"mean_absorption_fraction": 0.24073597329900212,
|
232 |
+
"full_absorption_rate": 0.07003891050583658,
|
233 |
+
"num_full_absorption": 18,
|
234 |
+
"num_probe_true_positives": 257,
|
235 |
+
"num_split_features": 1
|
236 |
+
}
|
237 |
+
],
|
238 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
239 |
+
"sae_lens_id": "custom_sae",
|
240 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3",
|
241 |
+
"sae_lens_version": "5.3.2",
|
242 |
+
"sae_cfg_dict": {
|
243 |
+
"model_name": "gemma-2-2b",
|
244 |
+
"d_in": 2304,
|
245 |
+
"d_sae": 16384,
|
246 |
+
"hook_layer": 12,
|
247 |
+
"hook_name": "blocks.12.hook_resid_post",
|
248 |
+
"context_size": null,
|
249 |
+
"hook_head_index": null,
|
250 |
+
"architecture": "matryoshka_batch_topk",
|
251 |
+
"apply_b_dec_to_input": null,
|
252 |
+
"finetuning_scaling_factor": null,
|
253 |
+
"activation_fn_str": "",
|
254 |
+
"prepend_bos": true,
|
255 |
+
"normalize_activations": "none",
|
256 |
+
"dtype": "bfloat16",
|
257 |
+
"device": "",
|
258 |
+
"dataset_path": "",
|
259 |
+
"dataset_trust_remote_code": true,
|
260 |
+
"seqpos_slice": [
|
261 |
+
null
|
262 |
+
],
|
263 |
+
"training_tokens": -100000,
|
264 |
+
"sae_lens_training_version": null,
|
265 |
+
"neuronpedia_id": null
|
266 |
+
},
|
267 |
+
"eval_result_unstructured": null
|
268 |
+
}
|
absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,268 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "absorption_first_letter",
|
3 |
+
"eval_config": {
|
4 |
+
"model_name": "gemma-2-2b",
|
5 |
+
"random_seed": 42,
|
6 |
+
"f1_jump_threshold": 0.03,
|
7 |
+
"max_k_value": 10,
|
8 |
+
"prompt_template": "{word} has the first letter:",
|
9 |
+
"prompt_token_pos": -6,
|
10 |
+
"llm_batch_size": 32,
|
11 |
+
"llm_dtype": "bfloat16",
|
12 |
+
"k_sparse_probe_l1_decay": 0.01,
|
13 |
+
"k_sparse_probe_batch_size": 4096,
|
14 |
+
"k_sparse_probe_num_epochs": 50
|
15 |
+
},
|
16 |
+
"eval_id": "81614709-a4e5-4ffc-a70e-787e336301fb",
|
17 |
+
"datetime_epoch_millis": 1737810824006,
|
18 |
+
"eval_result_metrics": {
|
19 |
+
"mean": {
|
20 |
+
"mean_absorption_fraction_score": 0.12344715765521998,
|
21 |
+
"mean_full_absorption_score": 0.020941627912473063,
|
22 |
+
"mean_num_split_features": 1.3076923076923077,
|
23 |
+
"std_dev_absorption_fraction_score": 0.22793801150350548,
|
24 |
+
"std_dev_full_absorption_score": 0.05550857979831386,
|
25 |
+
"std_dev_num_split_features": 0.5491251783869153
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"eval_result_details": [
|
29 |
+
{
|
30 |
+
"first_letter": "a",
|
31 |
+
"mean_absorption_fraction": 0.0010201140691341855,
|
32 |
+
"full_absorption_rate": 0.0,
|
33 |
+
"num_full_absorption": 0,
|
34 |
+
"num_probe_true_positives": 2565,
|
35 |
+
"num_split_features": 2
|
36 |
+
},
|
37 |
+
{
|
38 |
+
"first_letter": "b",
|
39 |
+
"mean_absorption_fraction": 0.0,
|
40 |
+
"full_absorption_rate": 0.0,
|
41 |
+
"num_full_absorption": 0,
|
42 |
+
"num_probe_true_positives": 1622,
|
43 |
+
"num_split_features": 1
|
44 |
+
},
|
45 |
+
{
|
46 |
+
"first_letter": "c",
|
47 |
+
"mean_absorption_fraction": 0.0,
|
48 |
+
"full_absorption_rate": 0.0,
|
49 |
+
"num_full_absorption": 0,
|
50 |
+
"num_probe_true_positives": 2798,
|
51 |
+
"num_split_features": 1
|
52 |
+
},
|
53 |
+
{
|
54 |
+
"first_letter": "d",
|
55 |
+
"mean_absorption_fraction": 0.0010590971415071488,
|
56 |
+
"full_absorption_rate": 0.0,
|
57 |
+
"num_full_absorption": 0,
|
58 |
+
"num_probe_true_positives": 1685,
|
59 |
+
"num_split_features": 1
|
60 |
+
},
|
61 |
+
{
|
62 |
+
"first_letter": "e",
|
63 |
+
"mean_absorption_fraction": 0.001095535100238956,
|
64 |
+
"full_absorption_rate": 0.0,
|
65 |
+
"num_full_absorption": 0,
|
66 |
+
"num_probe_true_positives": 1612,
|
67 |
+
"num_split_features": 2
|
68 |
+
},
|
69 |
+
{
|
70 |
+
"first_letter": "f",
|
71 |
+
"mean_absorption_fraction": 0.0,
|
72 |
+
"full_absorption_rate": 0.0,
|
73 |
+
"num_full_absorption": 0,
|
74 |
+
"num_probe_true_positives": 1232,
|
75 |
+
"num_split_features": 1
|
76 |
+
},
|
77 |
+
{
|
78 |
+
"first_letter": "g",
|
79 |
+
"mean_absorption_fraction": 0.008999780836552241,
|
80 |
+
"full_absorption_rate": 0.0,
|
81 |
+
"num_full_absorption": 0,
|
82 |
+
"num_probe_true_positives": 1063,
|
83 |
+
"num_split_features": 3
|
84 |
+
},
|
85 |
+
{
|
86 |
+
"first_letter": "h",
|
87 |
+
"mean_absorption_fraction": 0.0013575014784286386,
|
88 |
+
"full_absorption_rate": 0.0,
|
89 |
+
"num_full_absorption": 0,
|
90 |
+
"num_probe_true_positives": 1022,
|
91 |
+
"num_split_features": 1
|
92 |
+
},
|
93 |
+
{
|
94 |
+
"first_letter": "i",
|
95 |
+
"mean_absorption_fraction": 0.007083250387427594,
|
96 |
+
"full_absorption_rate": 0.0,
|
97 |
+
"num_full_absorption": 0,
|
98 |
+
"num_probe_true_positives": 1628,
|
99 |
+
"num_split_features": 2
|
100 |
+
},
|
101 |
+
{
|
102 |
+
"first_letter": "j",
|
103 |
+
"mean_absorption_fraction": 0.566561472742345,
|
104 |
+
"full_absorption_rate": 0.05251141552511415,
|
105 |
+
"num_full_absorption": 23,
|
106 |
+
"num_probe_true_positives": 438,
|
107 |
+
"num_split_features": 2
|
108 |
+
},
|
109 |
+
{
|
110 |
+
"first_letter": "k",
|
111 |
+
"mean_absorption_fraction": 0.7663905040703891,
|
112 |
+
"full_absorption_rate": 0.24495677233429394,
|
113 |
+
"num_full_absorption": 170,
|
114 |
+
"num_probe_true_positives": 694,
|
115 |
+
"num_split_features": 1
|
116 |
+
},
|
117 |
+
{
|
118 |
+
"first_letter": "l",
|
119 |
+
"mean_absorption_fraction": 0.00020750432331410754,
|
120 |
+
"full_absorption_rate": 0.0,
|
121 |
+
"num_full_absorption": 0,
|
122 |
+
"num_probe_true_positives": 1248,
|
123 |
+
"num_split_features": 1
|
124 |
+
},
|
125 |
+
{
|
126 |
+
"first_letter": "m",
|
127 |
+
"mean_absorption_fraction": 0.0024751152435083697,
|
128 |
+
"full_absorption_rate": 0.00340522133938706,
|
129 |
+
"num_full_absorption": 6,
|
130 |
+
"num_probe_true_positives": 1762,
|
131 |
+
"num_split_features": 1
|
132 |
+
},
|
133 |
+
{
|
134 |
+
"first_letter": "n",
|
135 |
+
"mean_absorption_fraction": 0.0,
|
136 |
+
"full_absorption_rate": 0.0,
|
137 |
+
"num_full_absorption": 0,
|
138 |
+
"num_probe_true_positives": 851,
|
139 |
+
"num_split_features": 1
|
140 |
+
},
|
141 |
+
{
|
142 |
+
"first_letter": "o",
|
143 |
+
"mean_absorption_fraction": 0.000556610817706727,
|
144 |
+
"full_absorption_rate": 0.0,
|
145 |
+
"num_full_absorption": 0,
|
146 |
+
"num_probe_true_positives": 1024,
|
147 |
+
"num_split_features": 2
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"first_letter": "p",
|
151 |
+
"mean_absorption_fraction": 0.0018693685060155129,
|
152 |
+
"full_absorption_rate": 0.0004317789291882556,
|
153 |
+
"num_full_absorption": 1,
|
154 |
+
"num_probe_true_positives": 2316,
|
155 |
+
"num_split_features": 1
|
156 |
+
},
|
157 |
+
{
|
158 |
+
"first_letter": "q",
|
159 |
+
"mean_absorption_fraction": 0.29103856296458736,
|
160 |
+
"full_absorption_rate": 0.016483516483516484,
|
161 |
+
"num_full_absorption": 3,
|
162 |
+
"num_probe_true_positives": 182,
|
163 |
+
"num_split_features": 1
|
164 |
+
},
|
165 |
+
{
|
166 |
+
"first_letter": "r",
|
167 |
+
"mean_absorption_fraction": 0.0,
|
168 |
+
"full_absorption_rate": 0.0,
|
169 |
+
"num_full_absorption": 0,
|
170 |
+
"num_probe_true_positives": 1642,
|
171 |
+
"num_split_features": 1
|
172 |
+
},
|
173 |
+
{
|
174 |
+
"first_letter": "s",
|
175 |
+
"mean_absorption_fraction": 7.926972035689905e-05,
|
176 |
+
"full_absorption_rate": 0.0,
|
177 |
+
"num_full_absorption": 0,
|
178 |
+
"num_probe_true_positives": 2784,
|
179 |
+
"num_split_features": 1
|
180 |
+
},
|
181 |
+
{
|
182 |
+
"first_letter": "t",
|
183 |
+
"mean_absorption_fraction": 0.0007885654451235668,
|
184 |
+
"full_absorption_rate": 0.0,
|
185 |
+
"num_full_absorption": 0,
|
186 |
+
"num_probe_true_positives": 1712,
|
187 |
+
"num_split_features": 1
|
188 |
+
},
|
189 |
+
{
|
190 |
+
"first_letter": "u",
|
191 |
+
"mean_absorption_fraction": 0.007405784661007207,
|
192 |
+
"full_absorption_rate": 0.0,
|
193 |
+
"num_full_absorption": 0,
|
194 |
+
"num_probe_true_positives": 787,
|
195 |
+
"num_split_features": 1
|
196 |
+
},
|
197 |
+
{
|
198 |
+
"first_letter": "v",
|
199 |
+
"mean_absorption_fraction": 0.0033170532123517386,
|
200 |
+
"full_absorption_rate": 0.0,
|
201 |
+
"num_full_absorption": 0,
|
202 |
+
"num_probe_true_positives": 771,
|
203 |
+
"num_split_features": 1
|
204 |
+
},
|
205 |
+
{
|
206 |
+
"first_letter": "w",
|
207 |
+
"mean_absorption_fraction": 0.23478386493978032,
|
208 |
+
"full_absorption_rate": 0.0028011204481792717,
|
209 |
+
"num_full_absorption": 2,
|
210 |
+
"num_probe_true_positives": 714,
|
211 |
+
"num_split_features": 1
|
212 |
+
},
|
213 |
+
{
|
214 |
+
"first_letter": "x",
|
215 |
+
"mean_absorption_fraction": 0.2196771993204489,
|
216 |
+
"full_absorption_rate": 0.0,
|
217 |
+
"num_full_absorption": 0,
|
218 |
+
"num_probe_true_positives": 101,
|
219 |
+
"num_split_features": 1
|
220 |
+
},
|
221 |
+
{
|
222 |
+
"first_letter": "y",
|
223 |
+
"mean_absorption_fraction": 0.5368367507064984,
|
224 |
+
"full_absorption_rate": 0.09937888198757763,
|
225 |
+
"num_full_absorption": 16,
|
226 |
+
"num_probe_true_positives": 161,
|
227 |
+
"num_split_features": 2
|
228 |
+
},
|
229 |
+
{
|
230 |
+
"first_letter": "z",
|
231 |
+
"mean_absorption_fraction": 0.5570231933489973,
|
232 |
+
"full_absorption_rate": 0.1245136186770428,
|
233 |
+
"num_full_absorption": 32,
|
234 |
+
"num_probe_true_positives": 257,
|
235 |
+
"num_split_features": 1
|
236 |
+
}
|
237 |
+
],
|
238 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
239 |
+
"sae_lens_id": "custom_sae",
|
240 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4",
|
241 |
+
"sae_lens_version": "5.3.2",
|
242 |
+
"sae_cfg_dict": {
|
243 |
+
"model_name": "gemma-2-2b",
|
244 |
+
"d_in": 2304,
|
245 |
+
"d_sae": 16384,
|
246 |
+
"hook_layer": 12,
|
247 |
+
"hook_name": "blocks.12.hook_resid_post",
|
248 |
+
"context_size": null,
|
249 |
+
"hook_head_index": null,
|
250 |
+
"architecture": "matryoshka_batch_topk",
|
251 |
+
"apply_b_dec_to_input": null,
|
252 |
+
"finetuning_scaling_factor": null,
|
253 |
+
"activation_fn_str": "",
|
254 |
+
"prepend_bos": true,
|
255 |
+
"normalize_activations": "none",
|
256 |
+
"dtype": "bfloat16",
|
257 |
+
"device": "",
|
258 |
+
"dataset_path": "",
|
259 |
+
"dataset_trust_remote_code": true,
|
260 |
+
"seqpos_slice": [
|
261 |
+
null
|
262 |
+
],
|
263 |
+
"training_tokens": -100000,
|
264 |
+
"sae_lens_training_version": null,
|
265 |
+
"neuronpedia_id": null
|
266 |
+
},
|
267 |
+
"eval_result_unstructured": null
|
268 |
+
}
|
absorption/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,268 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "absorption_first_letter",
|
3 |
+
"eval_config": {
|
4 |
+
"model_name": "gemma-2-2b",
|
5 |
+
"random_seed": 42,
|
6 |
+
"f1_jump_threshold": 0.03,
|
7 |
+
"max_k_value": 10,
|
8 |
+
"prompt_template": "{word} has the first letter:",
|
9 |
+
"prompt_token_pos": -6,
|
10 |
+
"llm_batch_size": 32,
|
11 |
+
"llm_dtype": "bfloat16",
|
12 |
+
"k_sparse_probe_l1_decay": 0.01,
|
13 |
+
"k_sparse_probe_batch_size": 4096,
|
14 |
+
"k_sparse_probe_num_epochs": 50
|
15 |
+
},
|
16 |
+
"eval_id": "1b835c25-d0bd-46a9-ba4a-65872e7977e3",
|
17 |
+
"datetime_epoch_millis": 1737813909106,
|
18 |
+
"eval_result_metrics": {
|
19 |
+
"mean": {
|
20 |
+
"mean_absorption_fraction_score": 0.20037975196337074,
|
21 |
+
"mean_full_absorption_score": 0.08445296532788284,
|
22 |
+
"mean_num_split_features": 1.2307692307692308,
|
23 |
+
"std_dev_absorption_fraction_score": 0.3571682950909613,
|
24 |
+
"std_dev_full_absorption_score": 0.18688583316919946,
|
25 |
+
"std_dev_num_split_features": 0.5144077999169381
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"eval_result_details": [
|
29 |
+
{
|
30 |
+
"first_letter": "a",
|
31 |
+
"mean_absorption_fraction": 0.0009566173313411924,
|
32 |
+
"full_absorption_rate": 0.0,
|
33 |
+
"num_full_absorption": 0,
|
34 |
+
"num_probe_true_positives": 2565,
|
35 |
+
"num_split_features": 2
|
36 |
+
},
|
37 |
+
{
|
38 |
+
"first_letter": "b",
|
39 |
+
"mean_absorption_fraction": 0.0002938069464259562,
|
40 |
+
"full_absorption_rate": 0.0,
|
41 |
+
"num_full_absorption": 0,
|
42 |
+
"num_probe_true_positives": 1622,
|
43 |
+
"num_split_features": 1
|
44 |
+
},
|
45 |
+
{
|
46 |
+
"first_letter": "c",
|
47 |
+
"mean_absorption_fraction": 8.607917666470771e-05,
|
48 |
+
"full_absorption_rate": 0.0,
|
49 |
+
"num_full_absorption": 0,
|
50 |
+
"num_probe_true_positives": 2798,
|
51 |
+
"num_split_features": 1
|
52 |
+
},
|
53 |
+
{
|
54 |
+
"first_letter": "d",
|
55 |
+
"mean_absorption_fraction": 0.0009619111121251727,
|
56 |
+
"full_absorption_rate": 0.0,
|
57 |
+
"num_full_absorption": 0,
|
58 |
+
"num_probe_true_positives": 1685,
|
59 |
+
"num_split_features": 1
|
60 |
+
},
|
61 |
+
{
|
62 |
+
"first_letter": "e",
|
63 |
+
"mean_absorption_fraction": 0.049988197876576865,
|
64 |
+
"full_absorption_rate": 0.0012406947890818859,
|
65 |
+
"num_full_absorption": 2,
|
66 |
+
"num_probe_true_positives": 1612,
|
67 |
+
"num_split_features": 1
|
68 |
+
},
|
69 |
+
{
|
70 |
+
"first_letter": "f",
|
71 |
+
"mean_absorption_fraction": 0.011091029140258684,
|
72 |
+
"full_absorption_rate": 0.0,
|
73 |
+
"num_full_absorption": 0,
|
74 |
+
"num_probe_true_positives": 1232,
|
75 |
+
"num_split_features": 1
|
76 |
+
},
|
77 |
+
{
|
78 |
+
"first_letter": "g",
|
79 |
+
"mean_absorption_fraction": 0.01330302514353583,
|
80 |
+
"full_absorption_rate": 0.0,
|
81 |
+
"num_full_absorption": 0,
|
82 |
+
"num_probe_true_positives": 1063,
|
83 |
+
"num_split_features": 1
|
84 |
+
},
|
85 |
+
{
|
86 |
+
"first_letter": "h",
|
87 |
+
"mean_absorption_fraction": 0.05030783748946356,
|
88 |
+
"full_absorption_rate": 0.0,
|
89 |
+
"num_full_absorption": 0,
|
90 |
+
"num_probe_true_positives": 1022,
|
91 |
+
"num_split_features": 1
|
92 |
+
},
|
93 |
+
{
|
94 |
+
"first_letter": "i",
|
95 |
+
"mean_absorption_fraction": 0.0030524798240067264,
|
96 |
+
"full_absorption_rate": 0.0,
|
97 |
+
"num_full_absorption": 0,
|
98 |
+
"num_probe_true_positives": 1628,
|
99 |
+
"num_split_features": 2
|
100 |
+
},
|
101 |
+
{
|
102 |
+
"first_letter": "j",
|
103 |
+
"mean_absorption_fraction": 0.9384884408833534,
|
104 |
+
"full_absorption_rate": 0.5844748858447488,
|
105 |
+
"num_full_absorption": 256,
|
106 |
+
"num_probe_true_positives": 438,
|
107 |
+
"num_split_features": 1
|
108 |
+
},
|
109 |
+
{
|
110 |
+
"first_letter": "k",
|
111 |
+
"mean_absorption_fraction": 0.9323589966108728,
|
112 |
+
"full_absorption_rate": 0.6051873198847262,
|
113 |
+
"num_full_absorption": 420,
|
114 |
+
"num_probe_true_positives": 694,
|
115 |
+
"num_split_features": 1
|
116 |
+
},
|
117 |
+
{
|
118 |
+
"first_letter": "l",
|
119 |
+
"mean_absorption_fraction": 0.002121626596979146,
|
120 |
+
"full_absorption_rate": 0.0,
|
121 |
+
"num_full_absorption": 0,
|
122 |
+
"num_probe_true_positives": 1248,
|
123 |
+
"num_split_features": 1
|
124 |
+
},
|
125 |
+
{
|
126 |
+
"first_letter": "m",
|
127 |
+
"mean_absorption_fraction": 0.0055798386979186015,
|
128 |
+
"full_absorption_rate": 0.00170261066969353,
|
129 |
+
"num_full_absorption": 3,
|
130 |
+
"num_probe_true_positives": 1762,
|
131 |
+
"num_split_features": 1
|
132 |
+
},
|
133 |
+
{
|
134 |
+
"first_letter": "n",
|
135 |
+
"mean_absorption_fraction": 0.0010513029126892014,
|
136 |
+
"full_absorption_rate": 0.0,
|
137 |
+
"num_full_absorption": 0,
|
138 |
+
"num_probe_true_positives": 851,
|
139 |
+
"num_split_features": 1
|
140 |
+
},
|
141 |
+
{
|
142 |
+
"first_letter": "o",
|
143 |
+
"mean_absorption_fraction": 0.00024869453245340525,
|
144 |
+
"full_absorption_rate": 0.0,
|
145 |
+
"num_full_absorption": 0,
|
146 |
+
"num_probe_true_positives": 1024,
|
147 |
+
"num_split_features": 3
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"first_letter": "p",
|
151 |
+
"mean_absorption_fraction": 0.0021469156020963694,
|
152 |
+
"full_absorption_rate": 0.0,
|
153 |
+
"num_full_absorption": 0,
|
154 |
+
"num_probe_true_positives": 2316,
|
155 |
+
"num_split_features": 1
|
156 |
+
},
|
157 |
+
{
|
158 |
+
"first_letter": "q",
|
159 |
+
"mean_absorption_fraction": 0.8798175401141061,
|
160 |
+
"full_absorption_rate": 0.38461538461538464,
|
161 |
+
"num_full_absorption": 70,
|
162 |
+
"num_probe_true_positives": 182,
|
163 |
+
"num_split_features": 1
|
164 |
+
},
|
165 |
+
{
|
166 |
+
"first_letter": "r",
|
167 |
+
"mean_absorption_fraction": 0.0,
|
168 |
+
"full_absorption_rate": 0.0,
|
169 |
+
"num_full_absorption": 0,
|
170 |
+
"num_probe_true_positives": 1642,
|
171 |
+
"num_split_features": 1
|
172 |
+
},
|
173 |
+
{
|
174 |
+
"first_letter": "s",
|
175 |
+
"mean_absorption_fraction": 6.947595948988776e-05,
|
176 |
+
"full_absorption_rate": 0.0,
|
177 |
+
"num_full_absorption": 0,
|
178 |
+
"num_probe_true_positives": 2784,
|
179 |
+
"num_split_features": 1
|
180 |
+
},
|
181 |
+
{
|
182 |
+
"first_letter": "t",
|
183 |
+
"mean_absorption_fraction": 0.0007906618601642995,
|
184 |
+
"full_absorption_rate": 0.0,
|
185 |
+
"num_full_absorption": 0,
|
186 |
+
"num_probe_true_positives": 1712,
|
187 |
+
"num_split_features": 1
|
188 |
+
},
|
189 |
+
{
|
190 |
+
"first_letter": "u",
|
191 |
+
"mean_absorption_fraction": 0.0014580450539102752,
|
192 |
+
"full_absorption_rate": 0.0,
|
193 |
+
"num_full_absorption": 0,
|
194 |
+
"num_probe_true_positives": 787,
|
195 |
+
"num_split_features": 2
|
196 |
+
},
|
197 |
+
{
|
198 |
+
"first_letter": "v",
|
199 |
+
"mean_absorption_fraction": 0.036502450497531785,
|
200 |
+
"full_absorption_rate": 0.0,
|
201 |
+
"num_full_absorption": 0,
|
202 |
+
"num_probe_true_positives": 771,
|
203 |
+
"num_split_features": 1
|
204 |
+
},
|
205 |
+
{
|
206 |
+
"first_letter": "w",
|
207 |
+
"mean_absorption_fraction": 0.031838181647245965,
|
208 |
+
"full_absorption_rate": 0.0,
|
209 |
+
"num_full_absorption": 0,
|
210 |
+
"num_probe_true_positives": 714,
|
211 |
+
"num_split_features": 2
|
212 |
+
},
|
213 |
+
{
|
214 |
+
"first_letter": "x",
|
215 |
+
"mean_absorption_fraction": 0.6415124978220123,
|
216 |
+
"full_absorption_rate": 0.0,
|
217 |
+
"num_full_absorption": 0,
|
218 |
+
"num_probe_true_positives": 101,
|
219 |
+
"num_split_features": 1
|
220 |
+
},
|
221 |
+
{
|
222 |
+
"first_letter": "y",
|
223 |
+
"mean_absorption_fraction": 0.8171887638571116,
|
224 |
+
"full_absorption_rate": 0.38509316770186336,
|
225 |
+
"num_full_absorption": 62,
|
226 |
+
"num_probe_true_positives": 161,
|
227 |
+
"num_split_features": 1
|
228 |
+
},
|
229 |
+
{
|
230 |
+
"first_letter": "z",
|
231 |
+
"mean_absorption_fraction": 0.7886591343593056,
|
232 |
+
"full_absorption_rate": 0.23346303501945526,
|
233 |
+
"num_full_absorption": 60,
|
234 |
+
"num_probe_true_positives": 257,
|
235 |
+
"num_split_features": 1
|
236 |
+
}
|
237 |
+
],
|
238 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
239 |
+
"sae_lens_id": "custom_sae",
|
240 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5",
|
241 |
+
"sae_lens_version": "5.3.2",
|
242 |
+
"sae_cfg_dict": {
|
243 |
+
"model_name": "gemma-2-2b",
|
244 |
+
"d_in": 2304,
|
245 |
+
"d_sae": 16384,
|
246 |
+
"hook_layer": 12,
|
247 |
+
"hook_name": "blocks.12.hook_resid_post",
|
248 |
+
"context_size": null,
|
249 |
+
"hook_head_index": null,
|
250 |
+
"architecture": "matryoshka_batch_topk",
|
251 |
+
"apply_b_dec_to_input": null,
|
252 |
+
"finetuning_scaling_factor": null,
|
253 |
+
"activation_fn_str": "",
|
254 |
+
"prepend_bos": true,
|
255 |
+
"normalize_activations": "none",
|
256 |
+
"dtype": "bfloat16",
|
257 |
+
"device": "",
|
258 |
+
"dataset_path": "",
|
259 |
+
"dataset_trust_remote_code": true,
|
260 |
+
"seqpos_slice": [
|
261 |
+
null
|
262 |
+
],
|
263 |
+
"training_tokens": -100000,
|
264 |
+
"sae_lens_training_version": null,
|
265 |
+
"neuronpedia_id": null
|
266 |
+
},
|
267 |
+
"eval_result_unstructured": null
|
268 |
+
}
|
autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "autointerp", "eval_config": {"model_name": "gemma-2-2b", "n_latents": 1000, "override_latents": null, "dead_latent_threshold": 15, "random_seed": 42, "dataset_name": "monology/pile-uncopyrighted", "llm_context_size": 128, "llm_batch_size": 32, "llm_dtype": "bfloat16", "buffer": 10, "no_overlap": true, "act_threshold_frac": 0.01, "total_tokens": 2000000, "scoring": true, "max_tokens_in_explanation": 30, "use_demos_in_explanation": true, "n_top_ex_for_generation": 10, "n_iw_sampled_ex_for_generation": 5, "n_top_ex_for_scoring": 2, "n_random_ex_for_scoring": 10, "n_iw_sampled_ex_for_scoring": 2}, "eval_id": "b0fe4c26-7ff2-4fd3-a4be-9635a8807d3f", "datetime_epoch_millis": 1737824874082, "eval_result_metrics": {"autointerp": {"autointerp_score": 0.8734756112098694, "autointerp_std_dev": 0.11937407404184341}}, "eval_result_details": [], "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}}
|
autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "autointerp", "eval_config": {"model_name": "gemma-2-2b", "n_latents": 1000, "override_latents": null, "dead_latent_threshold": 15, "random_seed": 42, "dataset_name": "monology/pile-uncopyrighted", "llm_context_size": 128, "llm_batch_size": 32, "llm_dtype": "bfloat16", "buffer": 10, "no_overlap": true, "act_threshold_frac": 0.01, "total_tokens": 2000000, "scoring": true, "max_tokens_in_explanation": 30, "use_demos_in_explanation": true, "n_top_ex_for_generation": 10, "n_iw_sampled_ex_for_generation": 5, "n_top_ex_for_scoring": 2, "n_random_ex_for_scoring": 10, "n_iw_sampled_ex_for_scoring": 2}, "eval_id": "2982dff5-2eab-41ab-9546-2f41c1f22010", "datetime_epoch_millis": 1737825408624, "eval_result_metrics": {"autointerp": {"autointerp_score": 0.8655674457550049, "autointerp_std_dev": 0.12434469908475876}}, "eval_result_details": [], "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}}
|
autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "autointerp", "eval_config": {"model_name": "gemma-2-2b", "n_latents": 1000, "override_latents": null, "dead_latent_threshold": 15, "random_seed": 42, "dataset_name": "monology/pile-uncopyrighted", "llm_context_size": 128, "llm_batch_size": 32, "llm_dtype": "bfloat16", "buffer": 10, "no_overlap": true, "act_threshold_frac": 0.01, "total_tokens": 2000000, "scoring": true, "max_tokens_in_explanation": 30, "use_demos_in_explanation": true, "n_top_ex_for_generation": 10, "n_iw_sampled_ex_for_generation": 5, "n_top_ex_for_scoring": 2, "n_random_ex_for_scoring": 10, "n_iw_sampled_ex_for_scoring": 2}, "eval_id": "4749767a-576c-49e4-aed6-d75e6eda37f4", "datetime_epoch_millis": 1737825937220, "eval_result_metrics": {"autointerp": {"autointerp_score": 0.8531865477561951, "autointerp_std_dev": 0.13071122765541077}}, "eval_result_details": [], "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}}
|
autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "autointerp", "eval_config": {"model_name": "gemma-2-2b", "n_latents": 1000, "override_latents": null, "dead_latent_threshold": 15, "random_seed": 42, "dataset_name": "monology/pile-uncopyrighted", "llm_context_size": 128, "llm_batch_size": 32, "llm_dtype": "bfloat16", "buffer": 10, "no_overlap": true, "act_threshold_frac": 0.01, "total_tokens": 2000000, "scoring": true, "max_tokens_in_explanation": 30, "use_demos_in_explanation": true, "n_top_ex_for_generation": 10, "n_iw_sampled_ex_for_generation": 5, "n_top_ex_for_scoring": 2, "n_random_ex_for_scoring": 10, "n_iw_sampled_ex_for_scoring": 2}, "eval_id": "d744a8a8-bfeb-4ea9-9bf4-2416946ce5fc", "datetime_epoch_millis": 1737826461659, "eval_result_metrics": {"autointerp": {"autointerp_score": 0.8406037092208862, "autointerp_std_dev": 0.1301453858613968}}, "eval_result_details": [], "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}}
|
autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "autointerp", "eval_config": {"model_name": "gemma-2-2b", "n_latents": 1000, "override_latents": null, "dead_latent_threshold": 15, "random_seed": 42, "dataset_name": "monology/pile-uncopyrighted", "llm_context_size": 128, "llm_batch_size": 32, "llm_dtype": "bfloat16", "buffer": 10, "no_overlap": true, "act_threshold_frac": 0.01, "total_tokens": 2000000, "scoring": true, "max_tokens_in_explanation": 30, "use_demos_in_explanation": true, "n_top_ex_for_generation": 10, "n_iw_sampled_ex_for_generation": 5, "n_top_ex_for_scoring": 2, "n_random_ex_for_scoring": 10, "n_iw_sampled_ex_for_scoring": 2}, "eval_id": "da22d218-d08f-43cb-af16-f306027353e8", "datetime_epoch_millis": 1737826999330, "eval_result_metrics": {"autointerp": {"autointerp_score": 0.8468455672264099, "autointerp_std_dev": 0.13139104843139648}}, "eval_result_details": [], "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}}
|
autointerp/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "autointerp", "eval_config": {"model_name": "gemma-2-2b", "n_latents": 1000, "override_latents": null, "dead_latent_threshold": 15, "random_seed": 42, "dataset_name": "monology/pile-uncopyrighted", "llm_context_size": 128, "llm_batch_size": 32, "llm_dtype": "bfloat16", "buffer": 10, "no_overlap": true, "act_threshold_frac": 0.01, "total_tokens": 2000000, "scoring": true, "max_tokens_in_explanation": 30, "use_demos_in_explanation": true, "n_top_ex_for_generation": 10, "n_iw_sampled_ex_for_generation": 5, "n_top_ex_for_scoring": 2, "n_random_ex_for_scoring": 10, "n_iw_sampled_ex_for_scoring": 2}, "eval_id": "1c3f0e1e-081d-4fb2-abf1-1ca43a34c7f5", "datetime_epoch_millis": 1737827536120, "eval_result_metrics": {"autointerp": {"autointerp_score": 0.8457461595535278, "autointerp_std_dev": 0.13389156758785248}}, "eval_result_details": [], "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}}
|
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:deada733df11a19fdc941a9ae1f6b25a66ff1b79c92e7f55b6720123e3c71db3
|
3 |
+
size 25784719
|
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:100e3e4989e2c178bbdeec508bc7000fef34956a159691709bf484833b82ea73
|
3 |
+
size 26205073
|
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:06e50b9bd79ab342347a3f58e06a8da1bc5e47d3652d20530375ac0aba1835ea
|
3 |
+
size 26411830
|
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a4d4ff4bce2129dcc65f1bbff9935303331482b92f7e66f3916364719c377316
|
3 |
+
size 26421008
|
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9169edf3f6fb6ba29b72cca1c7cd8a5183b25ba47e1e436018dc0f1dd498d973
|
3 |
+
size 26231137
|
autointerp_with_generations/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eb175481892acf92bc3fca2a1ab6417c7ace04335d56a6b830092da80fae971c
|
3 |
+
size 26250700
|
core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "core", "eval_config": {"model_name": "gemma-2-2b", "llm_dtype": "bfloat16", "batch_size_prompts": 16, "n_eval_reconstruction_batches": 200, "n_eval_sparsity_variance_batches": 2000, "dataset": "Skylion007/openwebtext", "context_size": 128, "compute_kl": true, "compute_ce_loss": true, "compute_l2_norms": true, "compute_sparsity_metrics": true, "compute_variance_metrics": true, "compute_featurewise_density_statistics": true, "compute_featurewise_weight_based_metrics": true, "exclude_special_tokens_from_reconstruction": true, "verbose": false}, "eval_id": "d3ea094f-c147-404e-9458-7bd05fd4f19b", "datetime_epoch_millis": 1737814432800, "eval_result_metrics": {"model_behavior_preservation": {"kl_div_score": 0.9642857142857143, "kl_div_with_ablation": 10.0625, "kl_div_with_sae": 0.359375}, "model_performance_preservation": {"ce_loss_score": 0.9638157894736842, "ce_loss_with_ablation": 12.4375, "ce_loss_with_sae": 3.28125, "ce_loss_without_sae": 2.9375}, "reconstruction_quality": {"explained_variance": 0.58984375, "mse": 2.53125, "cossim": 0.87109375}, "shrinkage": {"l2_norm_in": 149.0, "l2_norm_out": 133.0, "l2_ratio": 0.8828125, "relative_reconstruction_bias": 1.015625}, "sparsity": {"l0": 20.226354598999023, "l1": 308.0}, "token_stats": {"total_tokens_eval_reconstruction": 409600, "total_tokens_eval_sparsity_variance": 4096000}, "misc_metrics": {"freq_over_1_percent": 0.0206298828125, "freq_over_10_percent": 0.00042724609375, "normalized_freq_over_1_percent": 0.4437195360660553, "normalized_freq_over_10_percent": 0.09758249670267105, "average_max_encoder_cosine_sim": 0.17013800144195557, "average_max_decoder_cosine_sim": 0.233266681432724, "frac_alive": 0.9539794921875}}, "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}, "eval_result_unstructured": {}}
|
core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "core", "eval_config": {"model_name": "gemma-2-2b", "llm_dtype": "bfloat16", "batch_size_prompts": 16, "n_eval_reconstruction_batches": 200, "n_eval_sparsity_variance_batches": 2000, "dataset": "Skylion007/openwebtext", "context_size": 128, "compute_kl": true, "compute_ce_loss": true, "compute_l2_norms": true, "compute_sparsity_metrics": true, "compute_variance_metrics": true, "compute_featurewise_density_statistics": true, "compute_featurewise_weight_based_metrics": true, "exclude_special_tokens_from_reconstruction": true, "verbose": false}, "eval_id": "420bf07d-51a0-44f2-a65f-3020231a3f61", "datetime_epoch_millis": 1737814965476, "eval_result_metrics": {"model_behavior_preservation": {"kl_div_score": 0.9788431677018633, "kl_div_with_ablation": 10.0625, "kl_div_with_sae": 0.212890625}, "model_performance_preservation": {"ce_loss_score": 0.9786184210526315, "ce_loss_with_ablation": 12.4375, "ce_loss_with_sae": 3.140625, "ce_loss_without_sae": 2.9375}, "reconstruction_quality": {"explained_variance": 0.65234375, "mse": 2.15625, "cossim": 0.890625}, "shrinkage": {"l2_norm_in": 149.0, "l2_norm_out": 136.0, "l2_ratio": 0.90234375, "relative_reconstruction_bias": 1.015625}, "sparsity": {"l0": 40.54745864868164, "l1": 436.0}, "token_stats": {"total_tokens_eval_reconstruction": 409600, "total_tokens_eval_sparsity_variance": 4096000}, "misc_metrics": {"freq_over_1_percent": 0.05389404296875, "freq_over_10_percent": 0.00115966796875, "normalized_freq_over_1_percent": 0.5863255262374878, "normalized_freq_over_10_percent": 0.10462578386068344, "average_max_encoder_cosine_sim": 0.16614386439323425, "average_max_decoder_cosine_sim": 0.1930733323097229, "frac_alive": 0.9720458984375}}, "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}, "eval_result_unstructured": {}}
|
core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "core", "eval_config": {"model_name": "gemma-2-2b", "llm_dtype": "bfloat16", "batch_size_prompts": 16, "n_eval_reconstruction_batches": 200, "n_eval_sparsity_variance_batches": 2000, "dataset": "Skylion007/openwebtext", "context_size": 128, "compute_kl": true, "compute_ce_loss": true, "compute_l2_norms": true, "compute_sparsity_metrics": true, "compute_variance_metrics": true, "compute_featurewise_density_statistics": true, "compute_featurewise_weight_based_metrics": true, "exclude_special_tokens_from_reconstruction": true, "verbose": false}, "eval_id": "9d15e9c2-39b4-41bd-b7bb-2a6a74d33ce0", "datetime_epoch_millis": 1737815490780, "eval_result_metrics": {"model_behavior_preservation": {"kl_div_score": 0.9868012422360248, "kl_div_with_ablation": 10.0625, "kl_div_with_sae": 0.1328125}, "model_performance_preservation": {"ce_loss_score": 0.9868421052631579, "ce_loss_with_ablation": 12.4375, "ce_loss_with_sae": 3.0625, "ce_loss_without_sae": 2.9375}, "reconstruction_quality": {"explained_variance": 0.703125, "mse": 1.8203125, "cossim": 0.91015625}, "shrinkage": {"l2_norm_in": 149.0, "l2_norm_out": 138.0, "l2_ratio": 0.921875, "relative_reconstruction_bias": 1.015625}, "sparsity": {"l0": 81.85604858398438, "l1": 668.0}, "token_stats": {"total_tokens_eval_reconstruction": 409600, "total_tokens_eval_sparsity_variance": 4096000}, "misc_metrics": {"freq_over_1_percent": 0.1126708984375, "freq_over_10_percent": 0.00372314453125, "normalized_freq_over_1_percent": 0.7330043315887451, "normalized_freq_over_10_percent": 0.15342086553573608, "average_max_encoder_cosine_sim": 0.15199649333953857, "average_max_decoder_cosine_sim": 0.1583244800567627, "frac_alive": 0.9818115234375}}, "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}, "eval_result_unstructured": {}}
|
core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "core", "eval_config": {"model_name": "gemma-2-2b", "llm_dtype": "bfloat16", "batch_size_prompts": 16, "n_eval_reconstruction_batches": 200, "n_eval_sparsity_variance_batches": 2000, "dataset": "Skylion007/openwebtext", "context_size": 128, "compute_kl": true, "compute_ce_loss": true, "compute_l2_norms": true, "compute_sparsity_metrics": true, "compute_variance_metrics": true, "compute_featurewise_density_statistics": true, "compute_featurewise_weight_based_metrics": true, "exclude_special_tokens_from_reconstruction": true, "verbose": false}, "eval_id": "ab4df2e7-919f-479f-8c39-148fe06ae18a", "datetime_epoch_millis": 1737816013843, "eval_result_metrics": {"model_behavior_preservation": {"kl_div_score": 0.9914111024844721, "kl_div_with_ablation": 10.0625, "kl_div_with_sae": 0.08642578125}, "model_performance_preservation": {"ce_loss_score": 0.9917763157894737, "ce_loss_with_ablation": 12.4375, "ce_loss_with_sae": 3.015625, "ce_loss_without_sae": 2.9375}, "reconstruction_quality": {"explained_variance": 0.75390625, "mse": 1.515625, "cossim": 0.92578125}, "shrinkage": {"l2_norm_in": 149.0, "l2_norm_out": 140.0, "l2_ratio": 0.9375, "relative_reconstruction_bias": 1.0078125}, "sparsity": {"l0": 163.57395935058594, "l1": 1072.0}, "token_stats": {"total_tokens_eval_reconstruction": 409600, "total_tokens_eval_sparsity_variance": 4096000}, "misc_metrics": {"freq_over_1_percent": 0.1798095703125, "freq_over_10_percent": 0.01922607421875, "normalized_freq_over_1_percent": 0.8675092458724976, "normalized_freq_over_10_percent": 0.32524198293685913, "average_max_encoder_cosine_sim": 0.14852963387966156, "average_max_decoder_cosine_sim": 0.13293731212615967, "frac_alive": 0.97882080078125}}, "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}, "eval_result_unstructured": {}}
|
core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "core", "eval_config": {"model_name": "gemma-2-2b", "llm_dtype": "bfloat16", "batch_size_prompts": 16, "n_eval_reconstruction_batches": 200, "n_eval_sparsity_variance_batches": 2000, "dataset": "Skylion007/openwebtext", "context_size": 128, "compute_kl": true, "compute_ce_loss": true, "compute_l2_norms": true, "compute_sparsity_metrics": true, "compute_variance_metrics": true, "compute_featurewise_density_statistics": true, "compute_featurewise_weight_based_metrics": true, "exclude_special_tokens_from_reconstruction": true, "verbose": false}, "eval_id": "d66b6553-661e-4ed4-a1d2-d4f8b4639b53", "datetime_epoch_millis": 1737816542335, "eval_result_metrics": {"model_behavior_preservation": {"kl_div_score": 0.9945409549689441, "kl_div_with_ablation": 10.0625, "kl_div_with_sae": 0.054931640625}, "model_performance_preservation": {"ce_loss_score": 0.9950657894736842, "ce_loss_with_ablation": 12.4375, "ce_loss_with_sae": 2.984375, "ce_loss_without_sae": 2.9375}, "reconstruction_quality": {"explained_variance": 0.80078125, "mse": 1.1875, "cossim": 0.94140625}, "shrinkage": {"l2_norm_in": 149.0, "l2_norm_out": 143.0, "l2_ratio": 0.953125, "relative_reconstruction_bias": 1.0078125}, "sparsity": {"l0": 329.8325500488281, "l1": 1744.0}, "token_stats": {"total_tokens_eval_reconstruction": 409600, "total_tokens_eval_sparsity_variance": 4096000}, "misc_metrics": {"freq_over_1_percent": 0.190185546875, "freq_over_10_percent": 0.07415771484375, "normalized_freq_over_1_percent": 0.9459923505783081, "normalized_freq_over_10_percent": 0.7126327753067017, "average_max_encoder_cosine_sim": 0.1690574437379837, "average_max_decoder_cosine_sim": 0.11993047595024109, "frac_alive": 0.95184326171875}}, "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}, "eval_result_unstructured": {}}
|
core/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"eval_type_id": "core", "eval_config": {"model_name": "gemma-2-2b", "llm_dtype": "bfloat16", "batch_size_prompts": 16, "n_eval_reconstruction_batches": 200, "n_eval_sparsity_variance_batches": 2000, "dataset": "Skylion007/openwebtext", "context_size": 128, "compute_kl": true, "compute_ce_loss": true, "compute_l2_norms": true, "compute_sparsity_metrics": true, "compute_variance_metrics": true, "compute_featurewise_density_statistics": true, "compute_featurewise_weight_based_metrics": true, "exclude_special_tokens_from_reconstruction": true, "verbose": false}, "eval_id": "ad1fc031-2fd7-43a9-af62-b3cf85ec6fab", "datetime_epoch_millis": 1737817062836, "eval_result_metrics": {"model_behavior_preservation": {"kl_div_score": 0.9970035908385093, "kl_div_with_ablation": 10.0625, "kl_div_with_sae": 0.0301513671875}, "model_performance_preservation": {"ce_loss_score": 0.9967105263157895, "ce_loss_with_ablation": 12.4375, "ce_loss_with_sae": 2.96875, "ce_loss_without_sae": 2.9375}, "reconstruction_quality": {"explained_variance": 0.8671875, "mse": 0.765625, "cossim": 0.9609375}, "shrinkage": {"l2_norm_in": 149.0, "l2_norm_out": 145.0, "l2_ratio": 0.96484375, "relative_reconstruction_bias": 1.0078125}, "sparsity": {"l0": 668.866943359375, "l1": 2928.0}, "token_stats": {"total_tokens_eval_reconstruction": 409600, "total_tokens_eval_sparsity_variance": 4096000}, "misc_metrics": {"freq_over_1_percent": 0.160888671875, "freq_over_10_percent": 0.13446044921875, "normalized_freq_over_1_percent": 0.9867696166038513, "normalized_freq_over_10_percent": 0.9572452902793884, "average_max_encoder_cosine_sim": 0.21958830952644348, "average_max_decoder_cosine_sim": 0.11038094758987427, "frac_alive": 0.79705810546875}}, "sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764", "sae_lens_id": "custom_sae", "sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5", "sae_lens_version": "5.3.2", "sae_cfg_dict": {"model_name": "gemma-2-2b", "d_in": 2304, "d_sae": 16384, "hook_layer": 12, "hook_name": "blocks.12.hook_resid_post", "context_size": null, "hook_head_index": null, "architecture": "matryoshka_batch_topk", "apply_b_dec_to_input": null, "finetuning_scaling_factor": null, "activation_fn_str": "", "prepend_bos": true, "normalize_activations": "none", "dtype": "bfloat16", "device": "", "dataset_path": "", "dataset_trust_remote_code": true, "seqpos_slice": [null], "training_tokens": -100000, "sae_lens_training_version": null, "neuronpedia_id": null}, "eval_result_unstructured": {}}
|
core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
core_with_feature_statistics/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,323 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "scr",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
8 |
+
],
|
9 |
+
"perform_scr": true,
|
10 |
+
"early_stopping_patience": 20,
|
11 |
+
"train_set_size": 4000,
|
12 |
+
"test_set_size": 1000,
|
13 |
+
"context_length": 128,
|
14 |
+
"probe_train_batch_size": 16,
|
15 |
+
"probe_test_batch_size": 500,
|
16 |
+
"probe_epochs": 20,
|
17 |
+
"probe_lr": 0.001,
|
18 |
+
"probe_l1_penalty": 0.001,
|
19 |
+
"sae_batch_size": 125,
|
20 |
+
"llm_batch_size": 32,
|
21 |
+
"llm_dtype": "bfloat16",
|
22 |
+
"lower_vram_usage": false,
|
23 |
+
"model_name": "gemma-2-2b",
|
24 |
+
"n_values": [
|
25 |
+
2,
|
26 |
+
5,
|
27 |
+
10,
|
28 |
+
20,
|
29 |
+
50,
|
30 |
+
100,
|
31 |
+
500
|
32 |
+
],
|
33 |
+
"column1_vals_lookup": {
|
34 |
+
"LabHC/bias_in_bios_class_set1": [
|
35 |
+
[
|
36 |
+
"professor",
|
37 |
+
"nurse"
|
38 |
+
],
|
39 |
+
[
|
40 |
+
"architect",
|
41 |
+
"journalist"
|
42 |
+
],
|
43 |
+
[
|
44 |
+
"surgeon",
|
45 |
+
"psychologist"
|
46 |
+
],
|
47 |
+
[
|
48 |
+
"attorney",
|
49 |
+
"teacher"
|
50 |
+
]
|
51 |
+
],
|
52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
53 |
+
[
|
54 |
+
"Books",
|
55 |
+
"CDs_and_Vinyl"
|
56 |
+
],
|
57 |
+
[
|
58 |
+
"Software",
|
59 |
+
"Electronics"
|
60 |
+
],
|
61 |
+
[
|
62 |
+
"Pet_Supplies",
|
63 |
+
"Office_Products"
|
64 |
+
],
|
65 |
+
[
|
66 |
+
"Industrial_and_Scientific",
|
67 |
+
"Toys_and_Games"
|
68 |
+
]
|
69 |
+
]
|
70 |
+
}
|
71 |
+
},
|
72 |
+
"eval_id": "244d1947-a7d0-4a4d-8b8b-a5e9a9ba8427",
|
73 |
+
"datetime_epoch_millis": 1737818848829,
|
74 |
+
"eval_result_metrics": {
|
75 |
+
"scr_metrics": {
|
76 |
+
"scr_dir1_threshold_2": 0.18426428926142294,
|
77 |
+
"scr_metric_threshold_2": 0.17297445499200473,
|
78 |
+
"scr_dir2_threshold_2": 0.1719525753188406,
|
79 |
+
"scr_dir1_threshold_5": 0.11986111135412506,
|
80 |
+
"scr_metric_threshold_5": 0.2123169853274711,
|
81 |
+
"scr_dir2_threshold_5": 0.22951246399350642,
|
82 |
+
"scr_dir1_threshold_10": 0.08533140215956593,
|
83 |
+
"scr_metric_threshold_10": 0.23756152918820056,
|
84 |
+
"scr_dir2_threshold_10": 0.2681613529677045,
|
85 |
+
"scr_dir1_threshold_20": 0.05634989976488831,
|
86 |
+
"scr_metric_threshold_20": 0.2568280353363934,
|
87 |
+
"scr_dir2_threshold_20": 0.2802696748799438,
|
88 |
+
"scr_dir1_threshold_50": -0.17692548583841272,
|
89 |
+
"scr_metric_threshold_50": 0.2716830155948303,
|
90 |
+
"scr_dir2_threshold_50": 0.3075071205786852,
|
91 |
+
"scr_dir1_threshold_100": -0.254095228784124,
|
92 |
+
"scr_metric_threshold_100": 0.22231578730918813,
|
93 |
+
"scr_dir2_threshold_100": 0.2511809843173293,
|
94 |
+
"scr_dir1_threshold_500": -0.4827322571125042,
|
95 |
+
"scr_metric_threshold_500": 0.2099297153522725,
|
96 |
+
"scr_dir2_threshold_500": 0.23051268457311969
|
97 |
+
}
|
98 |
+
},
|
99 |
+
"eval_result_details": [
|
100 |
+
{
|
101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
|
102 |
+
"scr_dir1_threshold_2": 0.28125005820765137,
|
103 |
+
"scr_metric_threshold_2": 0.009852089465686957,
|
104 |
+
"scr_dir2_threshold_2": 0.009852089465686957,
|
105 |
+
"scr_dir1_threshold_5": 0.28125005820765137,
|
106 |
+
"scr_metric_threshold_5": 0.03201958438240929,
|
107 |
+
"scr_dir2_threshold_5": 0.03201958438240929,
|
108 |
+
"scr_dir1_threshold_10": 0.3125005820765137,
|
109 |
+
"scr_metric_threshold_10": 0.07635457421585395,
|
110 |
+
"scr_dir2_threshold_10": 0.07635457421585395,
|
111 |
+
"scr_dir1_threshold_20": 0.28125005820765137,
|
112 |
+
"scr_metric_threshold_20": 0.08374378812458251,
|
113 |
+
"scr_dir2_threshold_20": 0.08374378812458251,
|
114 |
+
"scr_dir1_threshold_50": 0.250000465661211,
|
115 |
+
"scr_metric_threshold_50": 0.024630517283144072,
|
116 |
+
"scr_dir2_threshold_50": 0.024630517283144072,
|
117 |
+
"scr_dir1_threshold_100": 0.20312514551912844,
|
118 |
+
"scr_metric_threshold_100": 0.051724056932709886,
|
119 |
+
"scr_dir2_threshold_100": 0.051724056932709886,
|
120 |
+
"scr_dir1_threshold_500": -0.23437473806556883,
|
121 |
+
"scr_metric_threshold_500": 0.10098509149899802,
|
122 |
+
"scr_dir2_threshold_500": 0.10098509149899802
|
123 |
+
},
|
124 |
+
{
|
125 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
|
126 |
+
"scr_dir1_threshold_2": 0.19626139590744457,
|
127 |
+
"scr_metric_threshold_2": 0.23919309396886884,
|
128 |
+
"scr_dir2_threshold_2": 0.23919309396886884,
|
129 |
+
"scr_dir1_threshold_5": 0.31775683754446676,
|
130 |
+
"scr_metric_threshold_5": 0.3170029204557986,
|
131 |
+
"scr_dir2_threshold_5": 0.3170029204557986,
|
132 |
+
"scr_dir1_threshold_10": 0.2056076276124383,
|
133 |
+
"scr_metric_threshold_10": 0.35446693007224356,
|
134 |
+
"scr_dir2_threshold_10": 0.35446693007224356,
|
135 |
+
"scr_dir1_threshold_20": 0.14018679094143036,
|
136 |
+
"scr_metric_threshold_20": 0.4063401477301967,
|
137 |
+
"scr_dir2_threshold_20": 0.4063401477301967,
|
138 |
+
"scr_dir1_threshold_50": -0.3271030692494605,
|
139 |
+
"scr_metric_threshold_50": 0.4063401477301967,
|
140 |
+
"scr_dir2_threshold_50": 0.4063401477301967,
|
141 |
+
"scr_dir1_threshold_100": -0.6915888371077373,
|
142 |
+
"scr_metric_threshold_100": 0.10086462806186652,
|
143 |
+
"scr_dir2_threshold_100": 0.10086462806186652,
|
144 |
+
"scr_dir1_threshold_500": -1.224299533969636,
|
145 |
+
"scr_metric_threshold_500": 0.1268012368908431,
|
146 |
+
"scr_dir2_threshold_500": 0.1268012368908431
|
147 |
+
},
|
148 |
+
{
|
149 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
|
150 |
+
"scr_dir1_threshold_2": 0.5156247962732202,
|
151 |
+
"scr_metric_threshold_2": 0.05316462152337083,
|
152 |
+
"scr_dir2_threshold_2": 0.05316462152337083,
|
153 |
+
"scr_dir1_threshold_5": 0.20312514551912844,
|
154 |
+
"scr_metric_threshold_5": 0.09113926495559606,
|
155 |
+
"scr_dir2_threshold_5": 0.09113926495559606,
|
156 |
+
"scr_dir1_threshold_10": 0.07812491268852294,
|
157 |
+
"scr_metric_threshold_10": 0.13924051014931188,
|
158 |
+
"scr_dir2_threshold_10": 0.13924051014931188,
|
159 |
+
"scr_dir1_threshold_20": 0.0937506402841651,
|
160 |
+
"scr_metric_threshold_20": 0.1721519281497015,
|
161 |
+
"scr_dir2_threshold_20": 0.1721519281497015,
|
162 |
+
"scr_dir1_threshold_50": -0.9531246798579175,
|
163 |
+
"scr_metric_threshold_50": 0.27088618215087035,
|
164 |
+
"scr_dir2_threshold_50": 0.27088618215087035,
|
165 |
+
"scr_dir1_threshold_100": -0.9687494761311376,
|
166 |
+
"scr_metric_threshold_100": 0.3164558146286684,
|
167 |
+
"scr_dir2_threshold_100": 0.3164558146286684,
|
168 |
+
"scr_dir1_threshold_500": -1.8593749708961742,
|
169 |
+
"scr_metric_threshold_500": 0.14936711191080249,
|
170 |
+
"scr_dir2_threshold_500": 0.14936711191080249
|
171 |
+
},
|
172 |
+
{
|
173 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
|
174 |
+
"scr_dir1_threshold_2": 0.14960621052073325,
|
175 |
+
"scr_metric_threshold_2": 0.16320476299460798,
|
176 |
+
"scr_dir2_threshold_2": 0.16320476299460798,
|
177 |
+
"scr_dir1_threshold_5": 0.17322841667108618,
|
178 |
+
"scr_metric_threshold_5": 0.2136499230091961,
|
179 |
+
"scr_dir2_threshold_5": 0.2136499230091961,
|
180 |
+
"scr_dir1_threshold_10": 0.21259844736981667,
|
181 |
+
"scr_metric_threshold_10": 0.249258302466854,
|
182 |
+
"scr_dir2_threshold_10": 0.249258302466854,
|
183 |
+
"scr_dir1_threshold_20": 0.08661444299943634,
|
184 |
+
"scr_metric_threshold_20": 0.14243334096225593,
|
185 |
+
"scr_dir2_threshold_20": 0.14243334096225593,
|
186 |
+
"scr_dir1_threshold_50": -0.08661397367164983,
|
187 |
+
"scr_metric_threshold_50": 0.17210690207611637,
|
188 |
+
"scr_dir2_threshold_50": 0.17210690207611637,
|
189 |
+
"scr_dir1_threshold_100": -0.23622018419238308,
|
190 |
+
"scr_metric_threshold_100": 0.008902139081508376,
|
191 |
+
"scr_dir2_threshold_100": 0.008902139081508376,
|
192 |
+
"scr_dir1_threshold_500": -0.19685015349365256,
|
193 |
+
"scr_metric_threshold_500": 0.03857570019536881,
|
194 |
+
"scr_dir2_threshold_500": 0.03857570019536881
|
195 |
+
},
|
196 |
+
{
|
197 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
|
198 |
+
"scr_dir1_threshold_2": 0.048913057562533044,
|
199 |
+
"scr_metric_threshold_2": 0.45882347441324484,
|
200 |
+
"scr_dir2_threshold_2": 0.45882347441324484,
|
201 |
+
"scr_dir1_threshold_5": -0.41304363318786347,
|
202 |
+
"scr_metric_threshold_5": 0.5058823941900663,
|
203 |
+
"scr_dir2_threshold_5": 0.5058823941900663,
|
204 |
+
"scr_dir1_threshold_10": -0.44021751806279735,
|
205 |
+
"scr_metric_threshold_10": 0.5490195620110374,
|
206 |
+
"scr_dir2_threshold_10": 0.5490195620110374,
|
207 |
+
"scr_dir1_threshold_20": -0.4293477697498678,
|
208 |
+
"scr_metric_threshold_20": 0.5843136934077262,
|
209 |
+
"scr_dir2_threshold_20": 0.5843136934077262,
|
210 |
+
"scr_dir1_threshold_50": -0.5489130575625331,
|
211 |
+
"scr_metric_threshold_50": 0.6156863065922739,
|
212 |
+
"scr_dir2_threshold_50": 0.6156863065922739,
|
213 |
+
"scr_dir1_threshold_100": -0.5326089210005287,
|
214 |
+
"scr_metric_threshold_100": 0.5568625984353196,
|
215 |
+
"scr_dir2_threshold_100": 0.5568625984353196,
|
216 |
+
"scr_dir1_threshold_500": -0.5217391726875992,
|
217 |
+
"scr_metric_threshold_500": 0.5294117372066224,
|
218 |
+
"scr_dir2_threshold_500": 0.5294117372066224
|
219 |
+
},
|
220 |
+
{
|
221 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
|
222 |
+
"scr_dir1_threshold_2": 0.11855670736578201,
|
223 |
+
"scr_metric_threshold_2": 0.1370969176230247,
|
224 |
+
"scr_dir2_threshold_2": 0.1370969176230247,
|
225 |
+
"scr_dir1_threshold_5": 0.12371128002507081,
|
226 |
+
"scr_metric_threshold_5": 0.1935485188968335,
|
227 |
+
"scr_dir2_threshold_5": 0.1935485188968335,
|
228 |
+
"scr_dir1_threshold_10": 0.015463717977866399,
|
229 |
+
"scr_metric_threshold_10": 0.23790326263293876,
|
230 |
+
"scr_dir2_threshold_10": 0.23790326263293876,
|
231 |
+
"scr_dir1_threshold_20": 0.08762865692919337,
|
232 |
+
"scr_metric_threshold_20": 0.2500001201706423,
|
233 |
+
"scr_dir2_threshold_20": 0.2500001201706423,
|
234 |
+
"scr_dir1_threshold_50": 0.0051545726592888,
|
235 |
+
"scr_metric_threshold_50": 0.2056451360932524,
|
236 |
+
"scr_dir2_threshold_50": 0.2056451360932524,
|
237 |
+
"scr_dir1_threshold_100": 0.0051545726592888,
|
238 |
+
"scr_metric_threshold_100": 0.282258006369044,
|
239 |
+
"scr_dir2_threshold_100": 0.282258006369044,
|
240 |
+
"scr_dir1_threshold_500": -0.05154664831417177,
|
241 |
+
"scr_metric_threshold_500": 0.3145161329087304,
|
242 |
+
"scr_dir2_threshold_500": 0.3145161329087304
|
243 |
+
},
|
244 |
+
{
|
245 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
|
246 |
+
"scr_dir1_threshold_2": 0.1081080573128186,
|
247 |
+
"scr_metric_threshold_2": 0.26666664900603304,
|
248 |
+
"scr_dir2_threshold_2": 0.26666664900603304,
|
249 |
+
"scr_dir1_threshold_5": 0.23423416892600485,
|
250 |
+
"scr_metric_threshold_5": 0.30666665960241324,
|
251 |
+
"scr_dir2_threshold_5": 0.30666665960241324,
|
252 |
+
"scr_dir1_threshold_10": 0.3243244404278432,
|
253 |
+
"scr_metric_threshold_10": 0.32000008477104136,
|
254 |
+
"scr_dir2_threshold_10": 0.32000008477104136,
|
255 |
+
"scr_dir1_threshold_20": 0.13063055806593132,
|
256 |
+
"scr_metric_threshold_20": 0.35555544370487596,
|
257 |
+
"scr_dir2_threshold_20": 0.35555544370487596,
|
258 |
+
"scr_dir1_threshold_50": 0.19369361387252446,
|
259 |
+
"scr_metric_threshold_50": 0.4266666913915537,
|
260 |
+
"scr_dir2_threshold_50": 0.4266666913915537,
|
261 |
+
"scr_dir1_threshold_100": 0.16666666666666666,
|
262 |
+
"scr_metric_threshold_100": 0.4399998516506776,
|
263 |
+
"scr_dir2_threshold_100": 0.4399998516506776,
|
264 |
+
"scr_dir1_threshold_500": 0.15765750527178912,
|
265 |
+
"scr_metric_threshold_500": 0.3511110569518347,
|
266 |
+
"scr_dir2_threshold_500": 0.3511110569518347
|
267 |
+
},
|
268 |
+
{
|
269 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
|
270 |
+
"scr_dir1_threshold_2": 0.05579403094120067,
|
271 |
+
"scr_metric_threshold_2": 0.05579403094120067,
|
272 |
+
"scr_dir2_threshold_2": 0.04761899355588758,
|
273 |
+
"scr_dir1_threshold_5": 0.03862661712745569,
|
274 |
+
"scr_metric_threshold_5": 0.03862661712745569,
|
275 |
+
"scr_dir2_threshold_5": 0.17619044645573817,
|
276 |
+
"scr_dir1_threshold_10": -0.025750992813676425,
|
277 |
+
"scr_metric_threshold_10": -0.025750992813676425,
|
278 |
+
"scr_dir2_threshold_10": 0.21904759742235502,
|
279 |
+
"scr_dir1_threshold_20": 0.060085820441166386,
|
280 |
+
"scr_metric_threshold_20": 0.060085820441166386,
|
281 |
+
"scr_dir2_threshold_20": 0.24761893678956953,
|
282 |
+
"scr_dir1_threshold_50": 0.05150224144123495,
|
283 |
+
"scr_metric_threshold_50": 0.05150224144123495,
|
284 |
+
"scr_dir2_threshold_50": 0.33809508131207394,
|
285 |
+
"scr_dir1_threshold_100": 0.021459203313710703,
|
286 |
+
"scr_metric_threshold_100": 0.021459203313710703,
|
287 |
+
"scr_dir2_threshold_100": 0.25238077937884024,
|
288 |
+
"scr_dir1_threshold_500": 0.06866965525497992,
|
289 |
+
"scr_metric_threshold_500": 0.06866965525497992,
|
290 |
+
"scr_dir2_threshold_500": 0.2333334090217574
|
291 |
+
}
|
292 |
+
],
|
293 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
294 |
+
"sae_lens_id": "custom_sae",
|
295 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0",
|
296 |
+
"sae_lens_version": "5.3.2",
|
297 |
+
"sae_cfg_dict": {
|
298 |
+
"model_name": "gemma-2-2b",
|
299 |
+
"d_in": 2304,
|
300 |
+
"d_sae": 16384,
|
301 |
+
"hook_layer": 12,
|
302 |
+
"hook_name": "blocks.12.hook_resid_post",
|
303 |
+
"context_size": null,
|
304 |
+
"hook_head_index": null,
|
305 |
+
"architecture": "matryoshka_batch_topk",
|
306 |
+
"apply_b_dec_to_input": null,
|
307 |
+
"finetuning_scaling_factor": null,
|
308 |
+
"activation_fn_str": "",
|
309 |
+
"prepend_bos": true,
|
310 |
+
"normalize_activations": "none",
|
311 |
+
"dtype": "bfloat16",
|
312 |
+
"device": "",
|
313 |
+
"dataset_path": "",
|
314 |
+
"dataset_trust_remote_code": true,
|
315 |
+
"seqpos_slice": [
|
316 |
+
null
|
317 |
+
],
|
318 |
+
"training_tokens": -100000,
|
319 |
+
"sae_lens_training_version": null,
|
320 |
+
"neuronpedia_id": null
|
321 |
+
},
|
322 |
+
"eval_result_unstructured": null
|
323 |
+
}
|
scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,323 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "scr",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
8 |
+
],
|
9 |
+
"perform_scr": true,
|
10 |
+
"early_stopping_patience": 20,
|
11 |
+
"train_set_size": 4000,
|
12 |
+
"test_set_size": 1000,
|
13 |
+
"context_length": 128,
|
14 |
+
"probe_train_batch_size": 16,
|
15 |
+
"probe_test_batch_size": 500,
|
16 |
+
"probe_epochs": 20,
|
17 |
+
"probe_lr": 0.001,
|
18 |
+
"probe_l1_penalty": 0.001,
|
19 |
+
"sae_batch_size": 125,
|
20 |
+
"llm_batch_size": 32,
|
21 |
+
"llm_dtype": "bfloat16",
|
22 |
+
"lower_vram_usage": false,
|
23 |
+
"model_name": "gemma-2-2b",
|
24 |
+
"n_values": [
|
25 |
+
2,
|
26 |
+
5,
|
27 |
+
10,
|
28 |
+
20,
|
29 |
+
50,
|
30 |
+
100,
|
31 |
+
500
|
32 |
+
],
|
33 |
+
"column1_vals_lookup": {
|
34 |
+
"LabHC/bias_in_bios_class_set1": [
|
35 |
+
[
|
36 |
+
"professor",
|
37 |
+
"nurse"
|
38 |
+
],
|
39 |
+
[
|
40 |
+
"architect",
|
41 |
+
"journalist"
|
42 |
+
],
|
43 |
+
[
|
44 |
+
"surgeon",
|
45 |
+
"psychologist"
|
46 |
+
],
|
47 |
+
[
|
48 |
+
"attorney",
|
49 |
+
"teacher"
|
50 |
+
]
|
51 |
+
],
|
52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
53 |
+
[
|
54 |
+
"Books",
|
55 |
+
"CDs_and_Vinyl"
|
56 |
+
],
|
57 |
+
[
|
58 |
+
"Software",
|
59 |
+
"Electronics"
|
60 |
+
],
|
61 |
+
[
|
62 |
+
"Pet_Supplies",
|
63 |
+
"Office_Products"
|
64 |
+
],
|
65 |
+
[
|
66 |
+
"Industrial_and_Scientific",
|
67 |
+
"Toys_and_Games"
|
68 |
+
]
|
69 |
+
]
|
70 |
+
}
|
71 |
+
},
|
72 |
+
"eval_id": "7c67dd45-96a6-43d4-8c07-6d9d65d05266",
|
73 |
+
"datetime_epoch_millis": 1737819337884,
|
74 |
+
"eval_result_metrics": {
|
75 |
+
"scr_metrics": {
|
76 |
+
"scr_dir1_threshold_2": 0.2056743018330487,
|
77 |
+
"scr_metric_threshold_2": 0.1620611500198696,
|
78 |
+
"scr_dir2_threshold_2": 0.16449320647201626,
|
79 |
+
"scr_dir1_threshold_5": 0.26996827195031564,
|
80 |
+
"scr_metric_threshold_5": 0.2378382259318422,
|
81 |
+
"scr_dir2_threshold_5": 0.2372787493948095,
|
82 |
+
"scr_dir1_threshold_10": 0.28239015818750307,
|
83 |
+
"scr_metric_threshold_10": 0.30700362695962724,
|
84 |
+
"scr_dir2_threshold_10": 0.3057237463667954,
|
85 |
+
"scr_dir1_threshold_20": 0.10131309170858085,
|
86 |
+
"scr_metric_threshold_20": 0.33196610334199106,
|
87 |
+
"scr_dir2_threshold_20": 0.33848304805299834,
|
88 |
+
"scr_dir1_threshold_50": 0.028058677683349006,
|
89 |
+
"scr_metric_threshold_50": 0.3493174310602765,
|
90 |
+
"scr_dir2_threshold_50": 0.34344425337337875,
|
91 |
+
"scr_dir1_threshold_100": -0.19765751631115763,
|
92 |
+
"scr_metric_threshold_100": 0.2818455109547832,
|
93 |
+
"scr_dir2_threshold_100": 0.2784197036352333,
|
94 |
+
"scr_dir1_threshold_500": -0.4114225428989186,
|
95 |
+
"scr_metric_threshold_500": 0.22262891090386205,
|
96 |
+
"scr_dir2_threshold_500": 0.24118346667031734
|
97 |
+
}
|
98 |
+
},
|
99 |
+
"eval_result_details": [
|
100 |
+
{
|
101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
|
102 |
+
"scr_dir1_threshold_2": 0.39062549476503666,
|
103 |
+
"scr_metric_threshold_2": 0.02709353964956581,
|
104 |
+
"scr_dir2_threshold_2": 0.02709353964956581,
|
105 |
+
"scr_dir1_threshold_5": 0.4062502910382569,
|
106 |
+
"scr_metric_threshold_5": 0.03448275355829437,
|
107 |
+
"scr_dir2_threshold_5": 0.03448275355829437,
|
108 |
+
"scr_dir1_threshold_10": 0.4531256111803394,
|
109 |
+
"scr_metric_threshold_10": 0.06896550711658873,
|
110 |
+
"scr_dir2_threshold_10": 0.06896550711658873,
|
111 |
+
"scr_dir1_threshold_20": 0.4531256111803394,
|
112 |
+
"scr_metric_threshold_20": 0.09852206913257629,
|
113 |
+
"scr_dir2_threshold_20": 0.09852206913257629,
|
114 |
+
"scr_dir1_threshold_50": 0.32812537834973393,
|
115 |
+
"scr_metric_threshold_50": 0.1403940365995992,
|
116 |
+
"scr_dir2_threshold_50": 0.1403940365995992,
|
117 |
+
"scr_dir1_threshold_100": 0.0937506402841651,
|
118 |
+
"scr_metric_threshold_100": 0.1970442850746159,
|
119 |
+
"scr_dir2_threshold_100": 0.1970442850746159,
|
120 |
+
"scr_dir1_threshold_500": -0.45312467985791743,
|
121 |
+
"scr_metric_threshold_500": 0.07635457421585395,
|
122 |
+
"scr_dir2_threshold_500": 0.07635457421585395
|
123 |
+
},
|
124 |
+
{
|
125 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
|
126 |
+
"scr_dir1_threshold_2": 0.25233655792624843,
|
127 |
+
"scr_metric_threshold_2": 0.3256485139892272,
|
128 |
+
"scr_dir2_threshold_2": 0.3256485139892272,
|
129 |
+
"scr_dir1_threshold_5": 0.31775683754446676,
|
130 |
+
"scr_metric_threshold_5": 0.35158512281820375,
|
131 |
+
"scr_dir2_threshold_5": 0.35158512281820375,
|
132 |
+
"scr_dir1_threshold_10": 0.2897198135878545,
|
133 |
+
"scr_metric_threshold_10": 0.41498556949231613,
|
134 |
+
"scr_dir2_threshold_10": 0.41498556949231613,
|
135 |
+
"scr_dir1_threshold_20": -1.0,
|
136 |
+
"scr_metric_threshold_20": 0.48703178147116627,
|
137 |
+
"scr_dir2_threshold_20": 0.48703178147116627,
|
138 |
+
"scr_dir1_threshold_50": -0.8691588837107738,
|
139 |
+
"scr_metric_threshold_50": 0.4034583404761569,
|
140 |
+
"scr_dir2_threshold_50": 0.4034583404761569,
|
141 |
+
"scr_dir1_threshold_100": -1.102804092332614,
|
142 |
+
"scr_metric_threshold_100": -0.04899141040391336,
|
143 |
+
"scr_dir2_threshold_100": -0.04899141040391336,
|
144 |
+
"scr_dir1_threshold_500": -1.2336452086218404,
|
145 |
+
"scr_metric_threshold_500": -0.07492784746158077,
|
146 |
+
"scr_dir2_threshold_500": -0.07492784746158077
|
147 |
+
},
|
148 |
+
{
|
149 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
|
150 |
+
"scr_dir1_threshold_2": 0.5625001164153027,
|
151 |
+
"scr_metric_threshold_2": 0.06835444871669703,
|
152 |
+
"scr_dir2_threshold_2": 0.06835444871669703,
|
153 |
+
"scr_dir1_threshold_5": 0.6093754365573852,
|
154 |
+
"scr_metric_threshold_5": 0.12911405928564068,
|
155 |
+
"scr_dir2_threshold_5": 0.12911405928564068,
|
156 |
+
"scr_dir1_threshold_10": 0.5625001164153027,
|
157 |
+
"scr_metric_threshold_10": 0.18227852991119212,
|
158 |
+
"scr_dir2_threshold_10": 0.18227852991119212,
|
159 |
+
"scr_dir1_threshold_20": 0.4531256111803394,
|
160 |
+
"scr_metric_threshold_20": 0.24303798958231634,
|
161 |
+
"scr_dir2_threshold_20": 0.24303798958231634,
|
162 |
+
"scr_dir1_threshold_50": 0.2968748544808716,
|
163 |
+
"scr_metric_threshold_50": 0.29113923477603215,
|
164 |
+
"scr_dir2_threshold_50": 0.29113923477603215,
|
165 |
+
"scr_dir1_threshold_100": -0.6249993015081835,
|
166 |
+
"scr_metric_threshold_100": 0.3088608255830956,
|
167 |
+
"scr_dir2_threshold_100": 0.3088608255830956,
|
168 |
+
"scr_dir1_threshold_500": -1.4843742724043578,
|
169 |
+
"scr_metric_threshold_500": 0.04556963247779803,
|
170 |
+
"scr_dir2_threshold_500": 0.04556963247779803
|
171 |
+
},
|
172 |
+
{
|
173 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
|
174 |
+
"scr_dir1_threshold_2": 0.14960621052073325,
|
175 |
+
"scr_metric_threshold_2": 0.0741840796530267,
|
176 |
+
"scr_dir2_threshold_2": 0.0741840796530267,
|
177 |
+
"scr_dir1_threshold_5": 0.18897671054725027,
|
178 |
+
"scr_metric_threshold_5": 0.21068260227148516,
|
179 |
+
"scr_dir2_threshold_5": 0.21068260227148516,
|
180 |
+
"scr_dir1_threshold_10": 0.14960621052073325,
|
181 |
+
"scr_metric_threshold_10": 0.28486650505613625,
|
182 |
+
"scr_dir2_threshold_10": 0.28486650505613625,
|
183 |
+
"scr_dir1_threshold_20": -0.09448788594583864,
|
184 |
+
"scr_metric_threshold_20": 0.23442152190992377,
|
185 |
+
"scr_dir2_threshold_20": 0.23442152190992377,
|
186 |
+
"scr_dir1_threshold_50": -0.04724394297291932,
|
187 |
+
"scr_metric_threshold_50": 0.30267078321915303,
|
188 |
+
"scr_dir2_threshold_50": 0.30267078321915303,
|
189 |
+
"scr_dir1_threshold_100": -0.16535403506911087,
|
190 |
+
"scr_metric_threshold_100": 0.11869442132381734,
|
191 |
+
"scr_dir2_threshold_100": 0.11869442132381734,
|
192 |
+
"scr_dir1_threshold_500": -0.14960621052073325,
|
193 |
+
"scr_metric_threshold_500": 0.050445160014588104,
|
194 |
+
"scr_dir2_threshold_500": 0.050445160014588104
|
195 |
+
},
|
196 |
+
{
|
197 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
|
198 |
+
"scr_dir1_threshold_2": 0.04347802143693829,
|
199 |
+
"scr_metric_threshold_2": 0.3529410802231785,
|
200 |
+
"scr_dir2_threshold_2": 0.3529410802231785,
|
201 |
+
"scr_dir1_threshold_5": 0.07608694243746696,
|
202 |
+
"scr_metric_threshold_5": 0.5176469488264898,
|
203 |
+
"scr_dir2_threshold_5": 0.5176469488264898,
|
204 |
+
"scr_dir1_threshold_10": 0.09782611512506609,
|
205 |
+
"scr_metric_threshold_10": 0.5921567298320084,
|
206 |
+
"scr_dir2_threshold_10": 0.5921567298320084,
|
207 |
+
"scr_dir1_threshold_20": 0.17934776974986782,
|
208 |
+
"scr_metric_threshold_20": 0.6196078248044149,
|
209 |
+
"scr_dir2_threshold_20": 0.6196078248044149,
|
210 |
+
"scr_dir1_threshold_50": -0.17934776974986782,
|
211 |
+
"scr_metric_threshold_50": 0.5882352116198672,
|
212 |
+
"scr_dir2_threshold_50": 0.5882352116198672,
|
213 |
+
"scr_dir1_threshold_100": -0.41847834537519824,
|
214 |
+
"scr_metric_threshold_100": 0.5764704232397345,
|
215 |
+
"scr_dir2_threshold_100": 0.5764704232397345,
|
216 |
+
"scr_dir1_threshold_500": -0.29891305756253306,
|
217 |
+
"scr_metric_threshold_500": 0.6078430364242823,
|
218 |
+
"scr_dir2_threshold_500": 0.6078430364242823
|
219 |
+
},
|
220 |
+
{
|
221 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
|
222 |
+
"scr_dir1_threshold_2": 0.11855670736578201,
|
223 |
+
"scr_metric_threshold_2": 0.07661287027579163,
|
224 |
+
"scr_dir2_threshold_2": 0.07661287027579163,
|
225 |
+
"scr_dir1_threshold_5": 0.25257713270943044,
|
226 |
+
"scr_metric_threshold_5": 0.1572581866250075,
|
227 |
+
"scr_dir2_threshold_5": 0.1572581866250075,
|
228 |
+
"scr_dir1_threshold_10": 0.2731957305870136,
|
229 |
+
"scr_metric_threshold_10": 0.22983885116865949,
|
230 |
+
"scr_dir2_threshold_10": 0.22983885116865949,
|
231 |
+
"scr_dir1_threshold_20": 0.2783503032463024,
|
232 |
+
"scr_metric_threshold_20": 0.29032265817460795,
|
233 |
+
"scr_dir2_threshold_20": 0.29032265817460795,
|
234 |
+
"scr_dir1_threshold_50": 0.2783503032463024,
|
235 |
+
"scr_metric_threshold_50": 0.2943548639067476,
|
236 |
+
"scr_dir2_threshold_50": 0.2943548639067476,
|
237 |
+
"scr_dir1_threshold_100": 0.26288658526843595,
|
238 |
+
"scr_metric_threshold_100": 0.3225807847142943,
|
239 |
+
"scr_dir2_threshold_100": 0.3225807847142943,
|
240 |
+
"scr_dir1_threshold_500": 0.2113399369542642,
|
241 |
+
"scr_metric_threshold_500": 0.4435483987261912,
|
242 |
+
"scr_dir2_threshold_500": 0.4435483987261912
|
243 |
+
},
|
244 |
+
{
|
245 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
|
246 |
+
"scr_dir1_threshold_2": 0.0810811101069608,
|
247 |
+
"scr_metric_threshold_2": 0.3244444715240827,
|
248 |
+
"scr_dir2_threshold_2": 0.3244444715240827,
|
249 |
+
"scr_dir1_threshold_5": 0.1756755595721568,
|
250 |
+
"scr_metric_threshold_5": 0.36888886887350414,
|
251 |
+
"scr_dir2_threshold_5": 0.36888886887350414,
|
252 |
+
"scr_dir1_threshold_10": 0.26576583107399515,
|
253 |
+
"scr_metric_threshold_10": 0.5155554860903967,
|
254 |
+
"scr_dir2_threshold_10": 0.5155554860903967,
|
255 |
+
"scr_dir1_threshold_20": 0.3693694419340687,
|
256 |
+
"scr_metric_threshold_20": 0.5111110993373553,
|
257 |
+
"scr_dir2_threshold_20": 0.5111110993373553,
|
258 |
+
"scr_dir1_threshold_50": 0.19369361387252446,
|
259 |
+
"scr_metric_threshold_50": 0.5511111099337356,
|
260 |
+
"scr_dir2_threshold_50": 0.5511111099337356,
|
261 |
+
"scr_dir1_threshold_100": 0.1891891674197794,
|
262 |
+
"scr_metric_threshold_100": 0.595555507283157,
|
263 |
+
"scr_dir2_threshold_100": 0.595555507283157,
|
264 |
+
"scr_dir1_threshold_500": 0.022522500753112733,
|
265 |
+
"scr_metric_threshold_500": 0.5377776847651073,
|
266 |
+
"scr_dir2_threshold_500": 0.5377776847651073
|
267 |
+
},
|
268 |
+
{
|
269 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
|
270 |
+
"scr_dir1_threshold_2": 0.047210196127387125,
|
271 |
+
"scr_metric_threshold_2": 0.047210196127387125,
|
272 |
+
"scr_dir2_threshold_2": 0.06666664774456064,
|
273 |
+
"scr_dir1_threshold_5": 0.13304726519611204,
|
274 |
+
"scr_metric_threshold_5": 0.13304726519611204,
|
275 |
+
"scr_dir2_threshold_5": 0.12857145289985059,
|
276 |
+
"scr_dir1_threshold_10": 0.1673818370097199,
|
277 |
+
"scr_metric_threshold_10": 0.1673818370097199,
|
278 |
+
"scr_dir2_threshold_10": 0.1571427922670651,
|
279 |
+
"scr_dir1_threshold_20": 0.17167388232356773,
|
280 |
+
"scr_metric_threshold_20": 0.17167388232356773,
|
281 |
+
"scr_dir2_threshold_20": 0.22380944001162575,
|
282 |
+
"scr_dir1_threshold_50": 0.22317586795092056,
|
283 |
+
"scr_metric_threshold_50": 0.22317586795092056,
|
284 |
+
"scr_dir2_threshold_50": 0.17619044645573817,
|
285 |
+
"scr_dir1_threshold_100": 0.18454925082346488,
|
286 |
+
"scr_metric_threshold_100": 0.18454925082346488,
|
287 |
+
"scr_dir2_threshold_100": 0.1571427922670651,
|
288 |
+
"scr_dir1_threshold_500": 0.09442064806865635,
|
289 |
+
"scr_metric_threshold_500": 0.09442064806865635,
|
290 |
+
"scr_dir2_threshold_500": 0.24285709420029883
|
291 |
+
}
|
292 |
+
],
|
293 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
294 |
+
"sae_lens_id": "custom_sae",
|
295 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1",
|
296 |
+
"sae_lens_version": "5.3.2",
|
297 |
+
"sae_cfg_dict": {
|
298 |
+
"model_name": "gemma-2-2b",
|
299 |
+
"d_in": 2304,
|
300 |
+
"d_sae": 16384,
|
301 |
+
"hook_layer": 12,
|
302 |
+
"hook_name": "blocks.12.hook_resid_post",
|
303 |
+
"context_size": null,
|
304 |
+
"hook_head_index": null,
|
305 |
+
"architecture": "matryoshka_batch_topk",
|
306 |
+
"apply_b_dec_to_input": null,
|
307 |
+
"finetuning_scaling_factor": null,
|
308 |
+
"activation_fn_str": "",
|
309 |
+
"prepend_bos": true,
|
310 |
+
"normalize_activations": "none",
|
311 |
+
"dtype": "bfloat16",
|
312 |
+
"device": "",
|
313 |
+
"dataset_path": "",
|
314 |
+
"dataset_trust_remote_code": true,
|
315 |
+
"seqpos_slice": [
|
316 |
+
null
|
317 |
+
],
|
318 |
+
"training_tokens": -100000,
|
319 |
+
"sae_lens_training_version": null,
|
320 |
+
"neuronpedia_id": null
|
321 |
+
},
|
322 |
+
"eval_result_unstructured": null
|
323 |
+
}
|
scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,323 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "scr",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
8 |
+
],
|
9 |
+
"perform_scr": true,
|
10 |
+
"early_stopping_patience": 20,
|
11 |
+
"train_set_size": 4000,
|
12 |
+
"test_set_size": 1000,
|
13 |
+
"context_length": 128,
|
14 |
+
"probe_train_batch_size": 16,
|
15 |
+
"probe_test_batch_size": 500,
|
16 |
+
"probe_epochs": 20,
|
17 |
+
"probe_lr": 0.001,
|
18 |
+
"probe_l1_penalty": 0.001,
|
19 |
+
"sae_batch_size": 125,
|
20 |
+
"llm_batch_size": 32,
|
21 |
+
"llm_dtype": "bfloat16",
|
22 |
+
"lower_vram_usage": false,
|
23 |
+
"model_name": "gemma-2-2b",
|
24 |
+
"n_values": [
|
25 |
+
2,
|
26 |
+
5,
|
27 |
+
10,
|
28 |
+
20,
|
29 |
+
50,
|
30 |
+
100,
|
31 |
+
500
|
32 |
+
],
|
33 |
+
"column1_vals_lookup": {
|
34 |
+
"LabHC/bias_in_bios_class_set1": [
|
35 |
+
[
|
36 |
+
"professor",
|
37 |
+
"nurse"
|
38 |
+
],
|
39 |
+
[
|
40 |
+
"architect",
|
41 |
+
"journalist"
|
42 |
+
],
|
43 |
+
[
|
44 |
+
"surgeon",
|
45 |
+
"psychologist"
|
46 |
+
],
|
47 |
+
[
|
48 |
+
"attorney",
|
49 |
+
"teacher"
|
50 |
+
]
|
51 |
+
],
|
52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
53 |
+
[
|
54 |
+
"Books",
|
55 |
+
"CDs_and_Vinyl"
|
56 |
+
],
|
57 |
+
[
|
58 |
+
"Software",
|
59 |
+
"Electronics"
|
60 |
+
],
|
61 |
+
[
|
62 |
+
"Pet_Supplies",
|
63 |
+
"Office_Products"
|
64 |
+
],
|
65 |
+
[
|
66 |
+
"Industrial_and_Scientific",
|
67 |
+
"Toys_and_Games"
|
68 |
+
]
|
69 |
+
]
|
70 |
+
}
|
71 |
+
},
|
72 |
+
"eval_id": "ca68edf9-e062-43a1-b394-bb596199a7c5",
|
73 |
+
"datetime_epoch_millis": 1737819796133,
|
74 |
+
"eval_result_metrics": {
|
75 |
+
"scr_metrics": {
|
76 |
+
"scr_dir1_threshold_2": 0.20964798636979887,
|
77 |
+
"scr_metric_threshold_2": 0.18380672681323953,
|
78 |
+
"scr_dir2_threshold_2": 0.18963393077779866,
|
79 |
+
"scr_dir1_threshold_5": 0.2602460032934394,
|
80 |
+
"scr_metric_threshold_5": 0.2691592535148927,
|
81 |
+
"scr_dir2_threshold_5": 0.27563278748167325,
|
82 |
+
"scr_dir1_threshold_10": 0.271486030529506,
|
83 |
+
"scr_metric_threshold_10": 0.3422087907224673,
|
84 |
+
"scr_dir2_threshold_10": 0.33045730939244783,
|
85 |
+
"scr_dir1_threshold_20": 0.29888980687911515,
|
86 |
+
"scr_metric_threshold_20": 0.2986418118515299,
|
87 |
+
"scr_dir2_threshold_20": 0.28860964337767203,
|
88 |
+
"scr_dir1_threshold_50": -0.05176562952223197,
|
89 |
+
"scr_metric_threshold_50": 0.3398816394465468,
|
90 |
+
"scr_dir2_threshold_50": 0.34199690636612745,
|
91 |
+
"scr_dir1_threshold_100": -0.5885617388243412,
|
92 |
+
"scr_metric_threshold_100": 0.2784789453128316,
|
93 |
+
"scr_dir2_threshold_100": 0.2865389294692976,
|
94 |
+
"scr_dir1_threshold_500": -1.1317925361499857,
|
95 |
+
"scr_metric_threshold_500": 0.19696019254085206,
|
96 |
+
"scr_dir2_threshold_500": 0.2462551236876233
|
97 |
+
}
|
98 |
+
},
|
99 |
+
"eval_result_details": [
|
100 |
+
{
|
101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
|
102 |
+
"scr_dir1_threshold_2": 0.4843752037267798,
|
103 |
+
"scr_metric_threshold_2": 0.022167494916722333,
|
104 |
+
"scr_dir2_threshold_2": 0.022167494916722333,
|
105 |
+
"scr_dir1_threshold_5": 0.5,
|
106 |
+
"scr_metric_threshold_5": 0.03201958438240929,
|
107 |
+
"scr_dir2_threshold_5": 0.03201958438240929,
|
108 |
+
"scr_dir1_threshold_10": 0.5,
|
109 |
+
"scr_metric_threshold_10": 0.07881774339173903,
|
110 |
+
"scr_dir2_threshold_10": 0.07881774339173903,
|
111 |
+
"scr_dir1_threshold_20": 0.5,
|
112 |
+
"scr_metric_threshold_20": 0.0566502484750167,
|
113 |
+
"scr_dir2_threshold_20": 0.0566502484750167,
|
114 |
+
"scr_dir1_threshold_50": -1.01562479627322,
|
115 |
+
"scr_metric_threshold_50": 0.19950730744103765,
|
116 |
+
"scr_dir2_threshold_50": 0.19950730744103765,
|
117 |
+
"scr_dir1_threshold_100": -1.1093745052349633,
|
118 |
+
"scr_metric_threshold_100": 0.27586202846635494,
|
119 |
+
"scr_dir2_threshold_100": 0.27586202846635494,
|
120 |
+
"scr_dir1_threshold_500": -2.1249993015081836,
|
121 |
+
"scr_metric_threshold_500": 0.07635457421585395,
|
122 |
+
"scr_dir2_threshold_500": 0.07635457421585395
|
123 |
+
},
|
124 |
+
{
|
125 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
|
126 |
+
"scr_dir1_threshold_2": 0.2242989769168465,
|
127 |
+
"scr_metric_threshold_2": 0.3170029204557986,
|
128 |
+
"scr_dir2_threshold_2": 0.3170029204557986,
|
129 |
+
"scr_dir1_threshold_5": 0.33644874390166457,
|
130 |
+
"scr_metric_threshold_5": 0.39481274694272833,
|
131 |
+
"scr_dir2_threshold_5": 0.39481274694272833,
|
132 |
+
"scr_dir1_threshold_10": 0.3831776742154747,
|
133 |
+
"scr_metric_threshold_10": 0.4726225734296581,
|
134 |
+
"scr_dir2_threshold_10": 0.4726225734296581,
|
135 |
+
"scr_dir1_threshold_20": 0.33644874390166457,
|
136 |
+
"scr_metric_threshold_20": 0.18443806905687588,
|
137 |
+
"scr_dir2_threshold_20": 0.18443806905687588,
|
138 |
+
"scr_dir1_threshold_50": -0.49532744120029276,
|
139 |
+
"scr_metric_threshold_50": 0.24495688024825762,
|
140 |
+
"scr_dir2_threshold_50": 0.24495688024825762,
|
141 |
+
"scr_dir1_threshold_100": -0.6355142321417231,
|
142 |
+
"scr_metric_threshold_100": -0.3170029204557986,
|
143 |
+
"scr_dir2_threshold_100": -0.3170029204557986,
|
144 |
+
"scr_dir1_threshold_500": -1.0,
|
145 |
+
"scr_metric_threshold_500": -0.34582133653881497,
|
146 |
+
"scr_dir2_threshold_500": -0.34582133653881497
|
147 |
+
},
|
148 |
+
{
|
149 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
|
150 |
+
"scr_dir1_threshold_2": 0.5468753201420825,
|
151 |
+
"scr_metric_threshold_2": 0.05063300880745302,
|
152 |
+
"scr_dir2_threshold_2": 0.05063300880745302,
|
153 |
+
"scr_dir1_threshold_5": 0.6250002328306055,
|
154 |
+
"scr_metric_threshold_5": 0.12151907024006786,
|
155 |
+
"scr_dir2_threshold_5": 0.12151907024006786,
|
156 |
+
"scr_dir1_threshold_10": 0.5,
|
157 |
+
"scr_metric_threshold_10": 0.21518994791158172,
|
158 |
+
"scr_dir2_threshold_10": 0.21518994791158172,
|
159 |
+
"scr_dir1_threshold_20": 0.4531256111803394,
|
160 |
+
"scr_metric_threshold_20": 0.15696210095637528,
|
161 |
+
"scr_dir2_threshold_20": 0.15696210095637528,
|
162 |
+
"scr_dir1_threshold_50": 0.03125052386886235,
|
163 |
+
"scr_metric_threshold_50": 0.1645570900019481,
|
164 |
+
"scr_dir2_threshold_50": 0.1645570900019481,
|
165 |
+
"scr_dir1_threshold_100": -3.7968739231584494,
|
166 |
+
"scr_metric_threshold_100": 0.19240513167268272,
|
167 |
+
"scr_dir2_threshold_100": 0.19240513167268272,
|
168 |
+
"scr_dir1_threshold_500": -4.874998835846973,
|
169 |
+
"scr_metric_threshold_500": -0.09113911405777667,
|
170 |
+
"scr_dir2_threshold_500": -0.09113911405777667
|
171 |
+
},
|
172 |
+
{
|
173 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
|
174 |
+
"scr_dir1_threshold_2": 0.1968506228214391,
|
175 |
+
"scr_metric_threshold_2": 0.16320476299460798,
|
176 |
+
"scr_dir2_threshold_2": 0.16320476299460798,
|
177 |
+
"scr_dir1_threshold_5": 0.12598447369816684,
|
178 |
+
"scr_metric_threshold_5": 0.2403561633853456,
|
179 |
+
"scr_dir2_threshold_5": 0.2403561633853456,
|
180 |
+
"scr_dir1_threshold_10": 0.007874381601975306,
|
181 |
+
"scr_metric_threshold_10": 0.31750756377608325,
|
182 |
+
"scr_dir2_threshold_10": 0.31750756377608325,
|
183 |
+
"scr_dir1_threshold_20": 0.06299223684908342,
|
184 |
+
"scr_metric_threshold_20": 0.1127597798483955,
|
185 |
+
"scr_dir2_threshold_20": 0.1127597798483955,
|
186 |
+
"scr_dir1_threshold_50": 0.299212890369253,
|
187 |
+
"scr_metric_threshold_50": 0.12166174206152826,
|
188 |
+
"scr_dir2_threshold_50": 0.12166174206152826,
|
189 |
+
"scr_dir1_threshold_100": 0.26771677194471133,
|
190 |
+
"scr_metric_threshold_100": -0.026706240376149513,
|
191 |
+
"scr_dir2_threshold_100": -0.026706240376149513,
|
192 |
+
"scr_dir1_threshold_500": -0.8425194078772914,
|
193 |
+
"scr_metric_threshold_500": -0.09792282242308967,
|
194 |
+
"scr_dir2_threshold_500": -0.09792282242308967
|
195 |
+
},
|
196 |
+
{
|
197 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
|
198 |
+
"scr_dir1_threshold_2": 0.021739172687599125,
|
199 |
+
"scr_metric_threshold_2": 0.3647058686033112,
|
200 |
+
"scr_dir2_threshold_2": 0.3647058686033112,
|
201 |
+
"scr_dir1_threshold_5": 0.05434776974986783,
|
202 |
+
"scr_metric_threshold_5": 0.5647058686033112,
|
203 |
+
"scr_dir2_threshold_5": 0.5647058686033112,
|
204 |
+
"scr_dir1_threshold_10": 0.10869553949973566,
|
205 |
+
"scr_metric_threshold_10": 0.6196078248044149,
|
206 |
+
"scr_dir2_threshold_10": 0.6196078248044149,
|
207 |
+
"scr_dir1_threshold_20": 0.08695636681213655,
|
208 |
+
"scr_metric_threshold_20": 0.6862745693856515,
|
209 |
+
"scr_dir2_threshold_20": 0.6862745693856515,
|
210 |
+
"scr_dir1_threshold_50": -0.04347834537519825,
|
211 |
+
"scr_metric_threshold_50": 0.6980391240220748,
|
212 |
+
"scr_dir2_threshold_50": 0.6980391240220748,
|
213 |
+
"scr_dir1_threshold_100": 0.04347802143693829,
|
214 |
+
"scr_metric_threshold_100": 0.7921567298320084,
|
215 |
+
"scr_dir2_threshold_100": 0.7921567298320084,
|
216 |
+
"scr_dir1_threshold_500": -0.2228261151250661,
|
217 |
+
"scr_metric_threshold_500": 0.7607843503911701,
|
218 |
+
"scr_dir2_threshold_500": 0.7607843503911701
|
219 |
+
},
|
220 |
+
{
|
221 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
|
222 |
+
"scr_dir1_threshold_2": 0.08762865692919337,
|
223 |
+
"scr_metric_threshold_2": 0.18951607282340924,
|
224 |
+
"scr_dir2_threshold_2": 0.18951607282340924,
|
225 |
+
"scr_dir1_threshold_5": 0.16494816853980915,
|
226 |
+
"scr_metric_threshold_5": 0.23387105690079912,
|
227 |
+
"scr_dir2_threshold_5": 0.23387105690079912,
|
228 |
+
"scr_dir1_threshold_10": 0.2061853642949754,
|
229 |
+
"scr_metric_threshold_10": 0.3145161329087304,
|
230 |
+
"scr_dir2_threshold_10": 0.3145161329087304,
|
231 |
+
"scr_dir1_threshold_20": 0.32989664432004623,
|
232 |
+
"scr_metric_threshold_20": 0.3629033227182599,
|
233 |
+
"scr_dir2_threshold_20": 0.3629033227182599,
|
234 |
+
"scr_dir1_threshold_50": 0.1958762189763978,
|
235 |
+
"scr_metric_threshold_50": 0.36693552845039956,
|
236 |
+
"scr_dir2_threshold_50": 0.36693552845039956,
|
237 |
+
"scr_dir1_threshold_100": 0.25257713270943044,
|
238 |
+
"scr_metric_threshold_100": 0.3225807847142943,
|
239 |
+
"scr_dir2_threshold_100": 0.3225807847142943,
|
240 |
+
"scr_dir1_threshold_500": 0.06701005905161024,
|
241 |
+
"scr_metric_threshold_500": 0.5604838070059485,
|
242 |
+
"scr_dir2_threshold_500": 0.5604838070059485
|
243 |
+
},
|
244 |
+
{
|
245 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
|
246 |
+
"scr_dir1_threshold_2": 0.0810811101069608,
|
247 |
+
"scr_metric_threshold_2": 0.32888885827712394,
|
248 |
+
"scr_dir2_threshold_2": 0.32888885827712394,
|
249 |
+
"scr_dir1_threshold_5": 0.19369361387252446,
|
250 |
+
"scr_metric_threshold_5": 0.48444451390960336,
|
251 |
+
"scr_dir2_threshold_5": 0.48444451390960336,
|
252 |
+
"scr_dir1_threshold_10": 0.3243244404278432,
|
253 |
+
"scr_metric_threshold_10": 0.5777776953614876,
|
254 |
+
"scr_dir2_threshold_10": 0.5777776953614876,
|
255 |
+
"scr_dir1_threshold_20": 0.44144139064615195,
|
256 |
+
"scr_metric_threshold_20": 0.6488889430481654,
|
257 |
+
"scr_dir2_threshold_20": 0.6488889430481654,
|
258 |
+
"scr_dir1_threshold_50": 0.4594594449465196,
|
259 |
+
"scr_metric_threshold_50": 0.7688889748373058,
|
260 |
+
"scr_dir2_threshold_50": 0.7688889748373058,
|
261 |
+
"scr_dir1_threshold_100": 0.07207194871208326,
|
262 |
+
"scr_metric_threshold_100": 0.7911111735120165,
|
263 |
+
"scr_dir2_threshold_100": 0.7911111735120165,
|
264 |
+
"scr_dir1_threshold_500": -0.004504446452745062,
|
265 |
+
"scr_metric_threshold_500": 0.7644443231747603,
|
266 |
+
"scr_dir2_threshold_500": 0.7644443231747603
|
267 |
+
},
|
268 |
+
{
|
269 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
|
270 |
+
"scr_dir1_threshold_2": 0.03433482762748996,
|
271 |
+
"scr_metric_threshold_2": 0.03433482762748996,
|
272 |
+
"scr_dir2_threshold_2": 0.08095245934396302,
|
273 |
+
"scr_dir1_threshold_5": 0.08154502375487709,
|
274 |
+
"scr_metric_threshold_5": 0.08154502375487709,
|
275 |
+
"scr_dir2_threshold_5": 0.1333332954891213,
|
276 |
+
"scr_dir1_threshold_10": 0.1416308441960435,
|
277 |
+
"scr_metric_threshold_10": 0.1416308441960435,
|
278 |
+
"scr_dir2_threshold_10": 0.04761899355588758,
|
279 |
+
"scr_dir1_threshold_20": 0.18025746132349915,
|
280 |
+
"scr_metric_threshold_20": 0.18025746132349915,
|
281 |
+
"scr_dir2_threshold_20": 0.10000011353263609,
|
282 |
+
"scr_dir1_threshold_50": 0.15450646850982275,
|
283 |
+
"scr_metric_threshold_50": 0.15450646850982275,
|
284 |
+
"scr_dir2_threshold_50": 0.17142860386646747,
|
285 |
+
"scr_dir1_threshold_100": 0.19742487513724416,
|
286 |
+
"scr_metric_threshold_100": 0.19742487513724416,
|
287 |
+
"scr_dir2_threshold_100": 0.2619047483889719,
|
288 |
+
"scr_dir1_threshold_500": -0.05150224144123495,
|
289 |
+
"scr_metric_threshold_500": -0.05150224144123495,
|
290 |
+
"scr_dir2_threshold_500": 0.34285720773293493
|
291 |
+
}
|
292 |
+
],
|
293 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
294 |
+
"sae_lens_id": "custom_sae",
|
295 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2",
|
296 |
+
"sae_lens_version": "5.3.2",
|
297 |
+
"sae_cfg_dict": {
|
298 |
+
"model_name": "gemma-2-2b",
|
299 |
+
"d_in": 2304,
|
300 |
+
"d_sae": 16384,
|
301 |
+
"hook_layer": 12,
|
302 |
+
"hook_name": "blocks.12.hook_resid_post",
|
303 |
+
"context_size": null,
|
304 |
+
"hook_head_index": null,
|
305 |
+
"architecture": "matryoshka_batch_topk",
|
306 |
+
"apply_b_dec_to_input": null,
|
307 |
+
"finetuning_scaling_factor": null,
|
308 |
+
"activation_fn_str": "",
|
309 |
+
"prepend_bos": true,
|
310 |
+
"normalize_activations": "none",
|
311 |
+
"dtype": "bfloat16",
|
312 |
+
"device": "",
|
313 |
+
"dataset_path": "",
|
314 |
+
"dataset_trust_remote_code": true,
|
315 |
+
"seqpos_slice": [
|
316 |
+
null
|
317 |
+
],
|
318 |
+
"training_tokens": -100000,
|
319 |
+
"sae_lens_training_version": null,
|
320 |
+
"neuronpedia_id": null
|
321 |
+
},
|
322 |
+
"eval_result_unstructured": null
|
323 |
+
}
|
scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,323 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "scr",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
8 |
+
],
|
9 |
+
"perform_scr": true,
|
10 |
+
"early_stopping_patience": 20,
|
11 |
+
"train_set_size": 4000,
|
12 |
+
"test_set_size": 1000,
|
13 |
+
"context_length": 128,
|
14 |
+
"probe_train_batch_size": 16,
|
15 |
+
"probe_test_batch_size": 500,
|
16 |
+
"probe_epochs": 20,
|
17 |
+
"probe_lr": 0.001,
|
18 |
+
"probe_l1_penalty": 0.001,
|
19 |
+
"sae_batch_size": 125,
|
20 |
+
"llm_batch_size": 32,
|
21 |
+
"llm_dtype": "bfloat16",
|
22 |
+
"lower_vram_usage": false,
|
23 |
+
"model_name": "gemma-2-2b",
|
24 |
+
"n_values": [
|
25 |
+
2,
|
26 |
+
5,
|
27 |
+
10,
|
28 |
+
20,
|
29 |
+
50,
|
30 |
+
100,
|
31 |
+
500
|
32 |
+
],
|
33 |
+
"column1_vals_lookup": {
|
34 |
+
"LabHC/bias_in_bios_class_set1": [
|
35 |
+
[
|
36 |
+
"professor",
|
37 |
+
"nurse"
|
38 |
+
],
|
39 |
+
[
|
40 |
+
"architect",
|
41 |
+
"journalist"
|
42 |
+
],
|
43 |
+
[
|
44 |
+
"surgeon",
|
45 |
+
"psychologist"
|
46 |
+
],
|
47 |
+
[
|
48 |
+
"attorney",
|
49 |
+
"teacher"
|
50 |
+
]
|
51 |
+
],
|
52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
53 |
+
[
|
54 |
+
"Books",
|
55 |
+
"CDs_and_Vinyl"
|
56 |
+
],
|
57 |
+
[
|
58 |
+
"Software",
|
59 |
+
"Electronics"
|
60 |
+
],
|
61 |
+
[
|
62 |
+
"Pet_Supplies",
|
63 |
+
"Office_Products"
|
64 |
+
],
|
65 |
+
[
|
66 |
+
"Industrial_and_Scientific",
|
67 |
+
"Toys_and_Games"
|
68 |
+
]
|
69 |
+
]
|
70 |
+
}
|
71 |
+
},
|
72 |
+
"eval_id": "5b15fa2b-d798-4f25-a0cc-1fea30ab2e16",
|
73 |
+
"datetime_epoch_millis": 1737820198682,
|
74 |
+
"eval_result_metrics": {
|
75 |
+
"scr_metrics": {
|
76 |
+
"scr_dir1_threshold_2": 0.25308338194638885,
|
77 |
+
"scr_metric_threshold_2": 0.17956576020976722,
|
78 |
+
"scr_dir2_threshold_2": 0.17520749316256728,
|
79 |
+
"scr_dir1_threshold_5": 0.30057075424163543,
|
80 |
+
"scr_metric_threshold_5": 0.2676904259128396,
|
81 |
+
"scr_dir2_threshold_5": 0.259326435137632,
|
82 |
+
"scr_dir1_threshold_10": 0.40949881637628427,
|
83 |
+
"scr_metric_threshold_10": 0.3520455920942588,
|
84 |
+
"scr_dir2_threshold_10": 0.33604570003176504,
|
85 |
+
"scr_dir1_threshold_20": 0.28210822334515906,
|
86 |
+
"scr_metric_threshold_20": 0.3157129000849953,
|
87 |
+
"scr_dir2_threshold_20": 0.2963689317141299,
|
88 |
+
"scr_dir1_threshold_50": -0.008304273656524384,
|
89 |
+
"scr_metric_threshold_50": 0.4047959507258682,
|
90 |
+
"scr_dir2_threshold_50": 0.40860752414985996,
|
91 |
+
"scr_dir1_threshold_100": -0.6375196205046844,
|
92 |
+
"scr_metric_threshold_100": 0.30790482128443697,
|
93 |
+
"scr_dir2_threshold_100": 0.34103121193887687,
|
94 |
+
"scr_dir1_threshold_500": -0.8367782681056807,
|
95 |
+
"scr_metric_threshold_500": 0.1197050315293546,
|
96 |
+
"scr_dir2_threshold_500": 0.15372813162703183
|
97 |
+
}
|
98 |
+
},
|
99 |
+
"eval_result_details": [
|
100 |
+
{
|
101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
|
102 |
+
"scr_dir1_threshold_2": 0.5156247962732202,
|
103 |
+
"scr_metric_threshold_2": 0.024630517283144072,
|
104 |
+
"scr_dir2_threshold_2": 0.024630517283144072,
|
105 |
+
"scr_dir1_threshold_5": 0.39062549476503666,
|
106 |
+
"scr_metric_threshold_5": 0.09852206913257629,
|
107 |
+
"scr_dir2_threshold_5": 0.09852206913257629,
|
108 |
+
"scr_dir1_threshold_10": 0.43749988358469727,
|
109 |
+
"scr_metric_threshold_10": 0.23891625254163884,
|
110 |
+
"scr_dir2_threshold_10": 0.23891625254163884,
|
111 |
+
"scr_dir1_threshold_20": -0.6249993015081835,
|
112 |
+
"scr_metric_threshold_20": 0.23152703863291027,
|
113 |
+
"scr_dir2_threshold_20": 0.23152703863291027,
|
114 |
+
"scr_dir1_threshold_50": -2.53124959254644,
|
115 |
+
"scr_metric_threshold_50": 0.22413782472418173,
|
116 |
+
"scr_dir2_threshold_50": 0.22413782472418173,
|
117 |
+
"scr_dir1_threshold_100": -3.578123981366101,
|
118 |
+
"scr_metric_threshold_100": -0.14039418340906254,
|
119 |
+
"scr_dir2_threshold_100": -0.14039418340906254,
|
120 |
+
"scr_dir1_threshold_500": -4.640624097781404,
|
121 |
+
"scr_metric_threshold_500": -0.19704443188407925,
|
122 |
+
"scr_dir2_threshold_500": -0.19704443188407925
|
123 |
+
},
|
124 |
+
{
|
125 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
|
126 |
+
"scr_dir1_threshold_2": 0.2336446515690506,
|
127 |
+
"scr_metric_threshold_2": 0.22190207867332085,
|
128 |
+
"scr_dir2_threshold_2": 0.22190207867332085,
|
129 |
+
"scr_dir1_threshold_5": 0.30841116289226267,
|
130 |
+
"scr_metric_threshold_5": 0.24783868750229743,
|
131 |
+
"scr_dir2_threshold_5": 0.24783868750229743,
|
132 |
+
"scr_dir1_threshold_10": 0.3831776742154747,
|
133 |
+
"scr_metric_threshold_10": 0.39193093968868853,
|
134 |
+
"scr_dir2_threshold_10": 0.39193093968868853,
|
135 |
+
"scr_dir1_threshold_20": 0.4299066045292848,
|
136 |
+
"scr_metric_threshold_20": 0.4726225734296581,
|
137 |
+
"scr_dir2_threshold_20": 0.4726225734296581,
|
138 |
+
"scr_dir1_threshold_50": 0.46728986019089086,
|
139 |
+
"scr_metric_threshold_50": 0.39769455419676814,
|
140 |
+
"scr_dir2_threshold_50": 0.39769455419676814,
|
141 |
+
"scr_dir1_threshold_100": -0.16822437195083229,
|
142 |
+
"scr_metric_threshold_100": 0.34005772203073537,
|
143 |
+
"scr_dir2_threshold_100": 0.34005772203073537,
|
144 |
+
"scr_dir1_threshold_500": -0.6635518131511251,
|
145 |
+
"scr_metric_threshold_500": 0.07204621197885014,
|
146 |
+
"scr_dir2_threshold_500": 0.07204621197885014
|
147 |
+
},
|
148 |
+
{
|
149 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
|
150 |
+
"scr_dir1_threshold_2": 0.5312505238688624,
|
151 |
+
"scr_metric_threshold_2": 0.07594943776226984,
|
152 |
+
"scr_dir2_threshold_2": 0.07594943776226984,
|
153 |
+
"scr_dir1_threshold_5": 0.5312505238688624,
|
154 |
+
"scr_metric_threshold_5": 0.09367087767151386,
|
155 |
+
"scr_dir2_threshold_5": 0.09367087767151386,
|
156 |
+
"scr_dir1_threshold_10": 0.5156247962732202,
|
157 |
+
"scr_metric_threshold_10": 0.15696210095637528,
|
158 |
+
"scr_dir2_threshold_10": 0.15696210095637528,
|
159 |
+
"scr_dir1_threshold_20": 0.4531256111803394,
|
160 |
+
"scr_metric_threshold_20": 0.25822796767346196,
|
161 |
+
"scr_dir2_threshold_20": 0.25822796767346196,
|
162 |
+
"scr_dir1_threshold_50": 0.20312514551912844,
|
163 |
+
"scr_metric_threshold_50": 0.43037974492534403,
|
164 |
+
"scr_dir2_threshold_50": 0.43037974492534403,
|
165 |
+
"scr_dir1_threshold_100": -0.7812500582076514,
|
166 |
+
"scr_metric_threshold_100": 0.42531651949350846,
|
167 |
+
"scr_dir2_threshold_100": 0.42531651949350846,
|
168 |
+
"scr_dir1_threshold_500": -0.17187462165026607,
|
169 |
+
"scr_metric_threshold_500": 0.1949367443886005,
|
170 |
+
"scr_dir2_threshold_500": 0.1949367443886005
|
171 |
+
},
|
172 |
+
{
|
173 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
|
174 |
+
"scr_dir1_threshold_2": 0.15748059212270857,
|
175 |
+
"scr_metric_threshold_2": 0.07121675891531577,
|
176 |
+
"scr_dir2_threshold_2": 0.07121675891531577,
|
177 |
+
"scr_dir1_threshold_5": 0.21259844736981667,
|
178 |
+
"scr_metric_threshold_5": 0.16617226060069454,
|
179 |
+
"scr_dir2_threshold_5": 0.16617226060069454,
|
180 |
+
"scr_dir1_threshold_10": 0.2755906842189001,
|
181 |
+
"scr_metric_threshold_10": 0.26112776228607326,
|
182 |
+
"scr_dir2_threshold_10": 0.26112776228607326,
|
183 |
+
"scr_dir1_threshold_20": 0.6377953421094501,
|
184 |
+
"scr_metric_threshold_20": -0.16320458612623237,
|
185 |
+
"scr_dir2_threshold_20": -0.16320458612623237,
|
186 |
+
"scr_dir1_threshold_50": 0.5433069868358249,
|
187 |
+
"scr_metric_threshold_50": -0.19584564484617933,
|
188 |
+
"scr_dir2_threshold_50": -0.19584564484617933,
|
189 |
+
"scr_dir1_threshold_100": -2.21259797804203,
|
190 |
+
"scr_metric_threshold_100": -0.3442136272838571,
|
191 |
+
"scr_dir2_threshold_100": -0.3442136272838571,
|
192 |
+
"scr_dir1_threshold_500": -2.472440368384766,
|
193 |
+
"scr_metric_threshold_500": -0.4183975300685082,
|
194 |
+
"scr_dir2_threshold_500": -0.4183975300685082
|
195 |
+
},
|
196 |
+
{
|
197 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
|
198 |
+
"scr_dir1_threshold_2": 0.021739172687599125,
|
199 |
+
"scr_metric_threshold_2": 0.5215684670386308,
|
200 |
+
"scr_dir2_threshold_2": 0.5215684670386308,
|
201 |
+
"scr_dir1_threshold_5": 0.0,
|
202 |
+
"scr_metric_threshold_5": 0.6352941313966888,
|
203 |
+
"scr_dir2_threshold_5": 0.6352941313966888,
|
204 |
+
"scr_dir1_threshold_10": 0.43478248193720265,
|
205 |
+
"scr_metric_threshold_10": 0.5686273868154523,
|
206 |
+
"scr_dir2_threshold_10": 0.5686273868154523,
|
207 |
+
"scr_dir1_threshold_20": 0.10326082731240087,
|
208 |
+
"scr_metric_threshold_20": 0.32156846703863085,
|
209 |
+
"scr_dir2_threshold_20": 0.32156846703863085,
|
210 |
+
"scr_dir1_threshold_50": -0.0815219785630617,
|
211 |
+
"scr_metric_threshold_50": 0.8156863065922738,
|
212 |
+
"scr_dir2_threshold_50": 0.8156863065922738,
|
213 |
+
"scr_dir1_threshold_100": 0.34239107899947135,
|
214 |
+
"scr_metric_threshold_100": 0.8666667445812365,
|
215 |
+
"scr_dir2_threshold_100": 0.8666667445812365,
|
216 |
+
"scr_dir1_threshold_500": 0.14673917268759912,
|
217 |
+
"scr_metric_threshold_500": 0.33725477363090467,
|
218 |
+
"scr_dir2_threshold_500": 0.33725477363090467
|
219 |
+
},
|
220 |
+
{
|
221 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
|
222 |
+
"scr_dir1_threshold_2": 0.25773170536871925,
|
223 |
+
"scr_metric_threshold_2": 0.0927419335456348,
|
224 |
+
"scr_dir2_threshold_2": 0.0927419335456348,
|
225 |
+
"scr_dir1_threshold_5": 0.3350515242197629,
|
226 |
+
"scr_metric_threshold_5": 0.2661291834404855,
|
227 |
+
"scr_dir2_threshold_5": 0.2661291834404855,
|
228 |
+
"scr_dir1_threshold_10": 0.4226801811489563,
|
229 |
+
"scr_metric_threshold_10": 0.31048392717659073,
|
230 |
+
"scr_dir2_threshold_10": 0.31048392717659073,
|
231 |
+
"scr_dir1_threshold_20": 0.3505152421976293,
|
232 |
+
"scr_metric_threshold_20": 0.40322586072222555,
|
233 |
+
"scr_dir2_threshold_20": 0.40322586072222555,
|
234 |
+
"scr_dir1_threshold_50": 0.3350515242197629,
|
235 |
+
"scr_metric_threshold_50": 0.608870996815478,
|
236 |
+
"scr_dir2_threshold_50": 0.608870996815478,
|
237 |
+
"scr_dir1_threshold_100": 0.38659786529350676,
|
238 |
+
"scr_metric_threshold_100": 0.6814516613591299,
|
239 |
+
"scr_dir2_threshold_100": 0.6814516613591299,
|
240 |
+
"scr_dir1_threshold_500": 0.3350515242197629,
|
241 |
+
"scr_metric_threshold_500": 0.6048387910833384,
|
242 |
+
"scr_dir2_threshold_500": 0.6048387910833384
|
243 |
+
},
|
244 |
+
{
|
245 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
|
246 |
+
"scr_dir1_threshold_2": 0.23423416892600485,
|
247 |
+
"scr_metric_threshold_2": 0.35555544370487596,
|
248 |
+
"scr_dir2_threshold_2": 0.35555544370487596,
|
249 |
+
"scr_dir1_threshold_5": 0.4549549984937745,
|
250 |
+
"scr_metric_threshold_5": 0.4622223152348926,
|
251 |
+
"scr_dir2_threshold_5": 0.4622223152348926,
|
252 |
+
"scr_dir1_threshold_10": 0.5405405550534804,
|
253 |
+
"scr_metric_threshold_10": 0.622222092710909,
|
254 |
+
"scr_dir2_threshold_10": 0.622222092710909,
|
255 |
+
"scr_dir1_threshold_20": 0.5810811101069608,
|
256 |
+
"scr_metric_threshold_20": 0.6755555284759174,
|
257 |
+
"scr_dir2_threshold_20": 0.6755555284759174,
|
258 |
+
"scr_dir1_threshold_50": 0.6756755595721567,
|
259 |
+
"scr_metric_threshold_50": 0.6355555178795371,
|
260 |
+
"scr_dir2_threshold_50": 0.6355555178795371,
|
261 |
+
"scr_dir1_threshold_100": 0.7522522232263725,
|
262 |
+
"scr_metric_threshold_100": 0.4755554754940165,
|
263 |
+
"scr_dir2_threshold_100": 0.4755554754940165,
|
264 |
+
"scr_dir1_threshold_500": 0.6351350045186764,
|
265 |
+
"scr_metric_threshold_500": 0.22666663840965287,
|
266 |
+
"scr_dir2_threshold_500": 0.22666663840965287
|
267 |
+
},
|
268 |
+
{
|
269 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
|
270 |
+
"scr_dir1_threshold_2": 0.07296144475494565,
|
271 |
+
"scr_metric_threshold_2": 0.07296144475494565,
|
272 |
+
"scr_dir2_threshold_2": 0.03809530837734615,
|
273 |
+
"scr_dir1_threshold_5": 0.17167388232356773,
|
274 |
+
"scr_metric_threshold_5": 0.17167388232356773,
|
275 |
+
"scr_dir2_threshold_5": 0.10476195612190681,
|
276 |
+
"scr_dir1_threshold_10": 0.266094274578342,
|
277 |
+
"scr_metric_threshold_10": 0.266094274578342,
|
278 |
+
"scr_dir2_threshold_10": 0.13809513807839202,
|
279 |
+
"scr_dir1_threshold_20": 0.3261803508333905,
|
280 |
+
"scr_metric_threshold_20": 0.3261803508333905,
|
281 |
+
"scr_dir2_threshold_20": 0.17142860386646747,
|
282 |
+
"scr_dir1_threshold_50": 0.3218883055195427,
|
283 |
+
"scr_metric_threshold_50": 0.3218883055195427,
|
284 |
+
"scr_dir2_threshold_50": 0.35238089291147634,
|
285 |
+
"scr_dir1_threshold_100": 0.15879825800978847,
|
286 |
+
"scr_metric_threshold_100": 0.15879825800978847,
|
287 |
+
"scr_dir2_threshold_100": 0.4238093832453077,
|
288 |
+
"scr_dir1_threshold_500": 0.13733905469607777,
|
289 |
+
"scr_metric_threshold_500": 0.13733905469607777,
|
290 |
+
"scr_dir2_threshold_500": 0.40952385547749554
|
291 |
+
}
|
292 |
+
],
|
293 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
294 |
+
"sae_lens_id": "custom_sae",
|
295 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3",
|
296 |
+
"sae_lens_version": "5.3.2",
|
297 |
+
"sae_cfg_dict": {
|
298 |
+
"model_name": "gemma-2-2b",
|
299 |
+
"d_in": 2304,
|
300 |
+
"d_sae": 16384,
|
301 |
+
"hook_layer": 12,
|
302 |
+
"hook_name": "blocks.12.hook_resid_post",
|
303 |
+
"context_size": null,
|
304 |
+
"hook_head_index": null,
|
305 |
+
"architecture": "matryoshka_batch_topk",
|
306 |
+
"apply_b_dec_to_input": null,
|
307 |
+
"finetuning_scaling_factor": null,
|
308 |
+
"activation_fn_str": "",
|
309 |
+
"prepend_bos": true,
|
310 |
+
"normalize_activations": "none",
|
311 |
+
"dtype": "bfloat16",
|
312 |
+
"device": "",
|
313 |
+
"dataset_path": "",
|
314 |
+
"dataset_trust_remote_code": true,
|
315 |
+
"seqpos_slice": [
|
316 |
+
null
|
317 |
+
],
|
318 |
+
"training_tokens": -100000,
|
319 |
+
"sae_lens_training_version": null,
|
320 |
+
"neuronpedia_id": null
|
321 |
+
},
|
322 |
+
"eval_result_unstructured": null
|
323 |
+
}
|
scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,323 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "scr",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
8 |
+
],
|
9 |
+
"perform_scr": true,
|
10 |
+
"early_stopping_patience": 20,
|
11 |
+
"train_set_size": 4000,
|
12 |
+
"test_set_size": 1000,
|
13 |
+
"context_length": 128,
|
14 |
+
"probe_train_batch_size": 16,
|
15 |
+
"probe_test_batch_size": 500,
|
16 |
+
"probe_epochs": 20,
|
17 |
+
"probe_lr": 0.001,
|
18 |
+
"probe_l1_penalty": 0.001,
|
19 |
+
"sae_batch_size": 125,
|
20 |
+
"llm_batch_size": 32,
|
21 |
+
"llm_dtype": "bfloat16",
|
22 |
+
"lower_vram_usage": false,
|
23 |
+
"model_name": "gemma-2-2b",
|
24 |
+
"n_values": [
|
25 |
+
2,
|
26 |
+
5,
|
27 |
+
10,
|
28 |
+
20,
|
29 |
+
50,
|
30 |
+
100,
|
31 |
+
500
|
32 |
+
],
|
33 |
+
"column1_vals_lookup": {
|
34 |
+
"LabHC/bias_in_bios_class_set1": [
|
35 |
+
[
|
36 |
+
"professor",
|
37 |
+
"nurse"
|
38 |
+
],
|
39 |
+
[
|
40 |
+
"architect",
|
41 |
+
"journalist"
|
42 |
+
],
|
43 |
+
[
|
44 |
+
"surgeon",
|
45 |
+
"psychologist"
|
46 |
+
],
|
47 |
+
[
|
48 |
+
"attorney",
|
49 |
+
"teacher"
|
50 |
+
]
|
51 |
+
],
|
52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
53 |
+
[
|
54 |
+
"Books",
|
55 |
+
"CDs_and_Vinyl"
|
56 |
+
],
|
57 |
+
[
|
58 |
+
"Software",
|
59 |
+
"Electronics"
|
60 |
+
],
|
61 |
+
[
|
62 |
+
"Pet_Supplies",
|
63 |
+
"Office_Products"
|
64 |
+
],
|
65 |
+
[
|
66 |
+
"Industrial_and_Scientific",
|
67 |
+
"Toys_and_Games"
|
68 |
+
]
|
69 |
+
]
|
70 |
+
}
|
71 |
+
},
|
72 |
+
"eval_id": "b2f01cfa-60b9-430e-b0b7-c0ed0f167cbd",
|
73 |
+
"datetime_epoch_millis": 1737820686480,
|
74 |
+
"eval_result_metrics": {
|
75 |
+
"scr_metrics": {
|
76 |
+
"scr_dir1_threshold_2": 0.2609642160050966,
|
77 |
+
"scr_metric_threshold_2": 0.1575305177334177,
|
78 |
+
"scr_dir2_threshold_2": 0.15603092100892096,
|
79 |
+
"scr_dir1_threshold_5": 0.35946648461237757,
|
80 |
+
"scr_metric_threshold_5": 0.24095131889802082,
|
81 |
+
"scr_dir2_threshold_5": 0.24885034355898503,
|
82 |
+
"scr_dir1_threshold_10": 0.22554907200923965,
|
83 |
+
"scr_metric_threshold_10": 0.32611325088452936,
|
84 |
+
"scr_dir2_threshold_10": 0.3310207745330495,
|
85 |
+
"scr_dir1_threshold_20": 0.22124937085415386,
|
86 |
+
"scr_metric_threshold_20": 0.3447237476871009,
|
87 |
+
"scr_dir2_threshold_20": 0.316584045212034,
|
88 |
+
"scr_dir1_threshold_50": -0.48821492044575865,
|
89 |
+
"scr_metric_threshold_50": 0.23019299810068877,
|
90 |
+
"scr_dir2_threshold_50": 0.21105599492439908,
|
91 |
+
"scr_dir1_threshold_100": -0.8624680469697904,
|
92 |
+
"scr_metric_threshold_100": 0.3554252166676827,
|
93 |
+
"scr_dir2_threshold_100": 0.31601943884826467,
|
94 |
+
"scr_dir1_threshold_500": -1.1291305729022583,
|
95 |
+
"scr_metric_threshold_500": 0.16562677969520906,
|
96 |
+
"scr_dir2_threshold_500": 0.12000550080541354
|
97 |
+
}
|
98 |
+
},
|
99 |
+
"eval_result_details": [
|
100 |
+
{
|
101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
|
102 |
+
"scr_dir1_threshold_2": 0.5312505238688624,
|
103 |
+
"scr_metric_threshold_2": 0.04187182065755959,
|
104 |
+
"scr_dir2_threshold_2": 0.04187182065755959,
|
105 |
+
"scr_dir1_threshold_5": 0.578124912688523,
|
106 |
+
"scr_metric_threshold_5": 0.051724056932709886,
|
107 |
+
"scr_dir2_threshold_5": 0.051724056932709886,
|
108 |
+
"scr_dir1_threshold_10": 0.42187508731147705,
|
109 |
+
"scr_metric_threshold_10": 0.1502462728747495,
|
110 |
+
"scr_dir2_threshold_10": 0.1502462728747495,
|
111 |
+
"scr_dir1_threshold_20": -0.031249592546440393,
|
112 |
+
"scr_metric_threshold_20": 0.19950730744103765,
|
113 |
+
"scr_dir2_threshold_20": 0.19950730744103765,
|
114 |
+
"scr_dir1_threshold_50": -1.906249359715835,
|
115 |
+
"scr_metric_threshold_50": 0.24876848881678912,
|
116 |
+
"scr_dir2_threshold_50": 0.24876848881678912,
|
117 |
+
"scr_dir1_threshold_100": -4.828124447027312,
|
118 |
+
"scr_metric_threshold_100": 0.2438422972744823,
|
119 |
+
"scr_dir2_threshold_100": 0.2438422972744823,
|
120 |
+
"scr_dir1_threshold_500": -4.53124959254644,
|
121 |
+
"scr_metric_threshold_500": -0.17241391460093516,
|
122 |
+
"scr_dir2_threshold_500": -0.17241391460093516
|
123 |
+
},
|
124 |
+
{
|
125 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
|
126 |
+
"scr_dir1_threshold_2": 0.31775683754446676,
|
127 |
+
"scr_metric_threshold_2": 0.11527383610337472,
|
128 |
+
"scr_dir2_threshold_2": 0.11527383610337472,
|
129 |
+
"scr_dir1_threshold_5": 0.49532688414750314,
|
130 |
+
"scr_metric_threshold_5": 0.2680115100518852,
|
131 |
+
"scr_dir2_threshold_5": 0.2680115100518852,
|
132 |
+
"scr_dir1_threshold_10": 0.4299066045292848,
|
133 |
+
"scr_metric_threshold_10": 0.33717591477669556,
|
134 |
+
"scr_dir2_threshold_10": 0.33717591477669556,
|
135 |
+
"scr_dir1_threshold_20": 0.5607477208185111,
|
136 |
+
"scr_metric_threshold_20": 0.3198847277098384,
|
137 |
+
"scr_dir2_threshold_20": 0.3198847277098384,
|
138 |
+
"scr_dir1_threshold_50": -0.7383177674215474,
|
139 |
+
"scr_metric_threshold_50": 0.38904896066333955,
|
140 |
+
"scr_dir2_threshold_50": 0.38904896066333955,
|
141 |
+
"scr_dir1_threshold_100": 0.17757004660303638,
|
142 |
+
"scr_metric_threshold_100": -0.19308349081899528,
|
143 |
+
"scr_dir2_threshold_100": -0.19308349081899528,
|
144 |
+
"scr_dir1_threshold_500": 0.551402046166307,
|
145 |
+
"scr_metric_threshold_500": -0.4092219549842365,
|
146 |
+
"scr_dir2_threshold_500": -0.4092219549842365
|
147 |
+
},
|
148 |
+
{
|
149 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
|
150 |
+
"scr_dir1_threshold_2": 0.5468753201420825,
|
151 |
+
"scr_metric_threshold_2": 0.06329122328486143,
|
152 |
+
"scr_dir2_threshold_2": 0.06329122328486143,
|
153 |
+
"scr_dir1_threshold_5": 0.5625001164153027,
|
154 |
+
"scr_metric_threshold_5": 0.1265822956719035,
|
155 |
+
"scr_dir2_threshold_5": 0.1265822956719035,
|
156 |
+
"scr_dir1_threshold_10": -1.0624991850928809,
|
157 |
+
"scr_metric_threshold_10": 0.1645570900019481,
|
158 |
+
"scr_dir2_threshold_10": 0.1645570900019481,
|
159 |
+
"scr_dir1_threshold_20": 0.1093754365573853,
|
160 |
+
"scr_metric_threshold_20": 0.1645570900019481,
|
161 |
+
"scr_dir2_threshold_20": 0.1645570900019481,
|
162 |
+
"scr_dir1_threshold_50": -2.328124447027312,
|
163 |
+
"scr_metric_threshold_50": 0.30126583653752276,
|
164 |
+
"scr_dir2_threshold_50": 0.30126583653752276,
|
165 |
+
"scr_dir1_threshold_100": -4.078123981366101,
|
166 |
+
"scr_metric_threshold_100": 0.3696202852542198,
|
167 |
+
"scr_dir2_threshold_100": 0.3696202852542198,
|
168 |
+
"scr_dir1_threshold_500": -3.999999068677578,
|
169 |
+
"scr_metric_threshold_500": 0.055696234239288635,
|
170 |
+
"scr_dir2_threshold_500": 0.055696234239288635
|
171 |
+
},
|
172 |
+
{
|
173 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
|
174 |
+
"scr_dir1_threshold_2": 0.2204723596440055,
|
175 |
+
"scr_metric_threshold_2": 0.0623146198338074,
|
176 |
+
"scr_dir2_threshold_2": 0.0623146198338074,
|
177 |
+
"scr_dir1_threshold_5": 0.43307080701382217,
|
178 |
+
"scr_metric_threshold_5": 0.14836798243767776,
|
179 |
+
"scr_dir2_threshold_5": 0.14836798243767776,
|
180 |
+
"scr_dir1_threshold_10": 0.7559054342056416,
|
181 |
+
"scr_metric_threshold_10": 0.23442152190992377,
|
182 |
+
"scr_dir2_threshold_10": 0.23442152190992377,
|
183 |
+
"scr_dir1_threshold_20": -0.2677163026169248,
|
184 |
+
"scr_metric_threshold_20": 0.2908013233999337,
|
185 |
+
"scr_dir2_threshold_20": 0.2908013233999337,
|
186 |
+
"scr_dir1_threshold_50": 0.30708680264344185,
|
187 |
+
"scr_metric_threshold_50": -0.029673561113860433,
|
188 |
+
"scr_dir2_threshold_50": -0.029673561113860433,
|
189 |
+
"scr_dir1_threshold_100": 0.35433074561636113,
|
190 |
+
"scr_metric_threshold_100": 0.4540060863945417,
|
191 |
+
"scr_dir2_threshold_100": 0.4540060863945417,
|
192 |
+
"scr_dir1_threshold_500": -0.4724408377125527,
|
193 |
+
"scr_metric_threshold_500": 0.25816026467998676,
|
194 |
+
"scr_dir2_threshold_500": 0.25816026467998676
|
195 |
+
},
|
196 |
+
{
|
197 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
|
198 |
+
"scr_dir1_threshold_2": -0.010869748312929542,
|
199 |
+
"scr_metric_threshold_2": 0.47058826279337757,
|
200 |
+
"scr_dir2_threshold_2": 0.47058826279337757,
|
201 |
+
"scr_dir1_threshold_5": 0.1684780214369383,
|
202 |
+
"scr_metric_threshold_5": 0.5764704232397345,
|
203 |
+
"scr_dir2_threshold_5": 0.5764704232397345,
|
204 |
+
"scr_dir1_threshold_10": 0.2717391726875991,
|
205 |
+
"scr_metric_threshold_10": 0.623529343016556,
|
206 |
+
"scr_dir2_threshold_10": 0.623529343016556,
|
207 |
+
"scr_dir1_threshold_20": 0.19565223025013218,
|
208 |
+
"scr_metric_threshold_20": 0.7450980437988962,
|
209 |
+
"scr_dir2_threshold_20": 0.7450980437988962,
|
210 |
+
"scr_dir1_threshold_50": 0.1630433092496035,
|
211 |
+
"scr_metric_threshold_50": -0.27058826279337755,
|
212 |
+
"scr_dir2_threshold_50": -0.27058826279337755,
|
213 |
+
"scr_dir1_threshold_100": 0.3152171941245374,
|
214 |
+
"scr_metric_threshold_100": 0.5372547736309047,
|
215 |
+
"scr_dir2_threshold_100": 0.5372547736309047,
|
216 |
+
"scr_dir1_threshold_500": -1.3858697483129296,
|
217 |
+
"scr_metric_threshold_500": 0.10196064223421587,
|
218 |
+
"scr_dir2_threshold_500": 0.10196064223421587
|
219 |
+
},
|
220 |
+
{
|
221 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
|
222 |
+
"scr_dir1_threshold_2": 0.17010304843952587,
|
223 |
+
"scr_metric_threshold_2": 0.1491935348194436,
|
224 |
+
"scr_dir2_threshold_2": 0.1491935348194436,
|
225 |
+
"scr_dir1_threshold_5": 0.2061853642949754,
|
226 |
+
"scr_metric_threshold_5": 0.21774199363095595,
|
227 |
+
"scr_dir2_threshold_5": 0.21774199363095595,
|
228 |
+
"scr_dir1_threshold_10": 0.3659792674159237,
|
229 |
+
"scr_metric_threshold_10": 0.3588711169861203,
|
230 |
+
"scr_dir2_threshold_10": 0.3588711169861203,
|
231 |
+
"scr_dir1_threshold_20": 0.45876280424483373,
|
232 |
+
"scr_metric_threshold_20": 0.23790326263293876,
|
233 |
+
"scr_dir2_threshold_20": 0.23790326263293876,
|
234 |
+
"scr_dir1_threshold_50": 0.190721646317109,
|
235 |
+
"scr_metric_threshold_50": 0.3064517214444511,
|
236 |
+
"scr_dir2_threshold_50": 0.3064517214444511,
|
237 |
+
"scr_dir1_threshold_100": 0.3969070106120844,
|
238 |
+
"scr_metric_threshold_100": 0.31048392717659073,
|
239 |
+
"scr_dir2_threshold_100": 0.31048392717659073,
|
240 |
+
"scr_dir1_threshold_500": 0.190721646317109,
|
241 |
+
"scr_metric_threshold_500": 0.6693548038214264,
|
242 |
+
"scr_dir2_threshold_500": 0.6693548038214264
|
243 |
+
},
|
244 |
+
{
|
245 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
|
246 |
+
"scr_dir1_threshold_2": 0.24774777677362747,
|
247 |
+
"scr_metric_threshold_2": 0.29333323443378506,
|
248 |
+
"scr_dir2_threshold_2": 0.29333323443378506,
|
249 |
+
"scr_dir1_threshold_5": 0.3333333333333333,
|
250 |
+
"scr_metric_threshold_5": 0.4399998516506776,
|
251 |
+
"scr_dir2_threshold_5": 0.4399998516506776,
|
252 |
+
"scr_dir1_threshold_10": 0.43693694419340684,
|
253 |
+
"scr_metric_threshold_10": 0.5555554966867768,
|
254 |
+
"scr_dir2_threshold_10": 0.5555554966867768,
|
255 |
+
"scr_dir1_threshold_20": 0.49549555354725494,
|
256 |
+
"scr_metric_threshold_20": 0.5511111099337356,
|
257 |
+
"scr_dir2_threshold_20": 0.5511111099337356,
|
258 |
+
"scr_dir1_threshold_50": 0.23873861537874994,
|
259 |
+
"scr_metric_threshold_50": 0.7288889642409256,
|
260 |
+
"scr_dir2_threshold_50": 0.7288889642409256,
|
261 |
+
"scr_dir1_threshold_100": 0.4234233363457843,
|
262 |
+
"scr_metric_threshold_100": 0.7822221350964297,
|
263 |
+
"scr_dir2_threshold_100": 0.7822221350964297,
|
264 |
+
"scr_dir1_threshold_500": 0.4684683378520097,
|
265 |
+
"scr_metric_threshold_500": 0.6755555284759174,
|
266 |
+
"scr_dir2_threshold_500": 0.6755555284759174
|
267 |
+
},
|
268 |
+
{
|
269 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
|
270 |
+
"scr_dir1_threshold_2": 0.0643776099411321,
|
271 |
+
"scr_metric_threshold_2": 0.0643776099411321,
|
272 |
+
"scr_dir2_threshold_2": 0.05238083614515829,
|
273 |
+
"scr_dir1_threshold_5": 0.09871243756862208,
|
274 |
+
"scr_metric_threshold_5": 0.09871243756862208,
|
275 |
+
"scr_dir2_threshold_5": 0.1619046348563358,
|
276 |
+
"scr_dir1_threshold_10": 0.18454925082346488,
|
277 |
+
"scr_metric_threshold_10": 0.18454925082346488,
|
278 |
+
"scr_dir2_threshold_10": 0.22380944001162575,
|
279 |
+
"scr_dir1_threshold_20": 0.2489271165784791,
|
280 |
+
"scr_metric_threshold_20": 0.2489271165784791,
|
281 |
+
"scr_dir2_threshold_20": 0.02380949677794379,
|
282 |
+
"scr_dir1_threshold_50": 0.1673818370097199,
|
283 |
+
"scr_metric_threshold_50": 0.1673818370097199,
|
284 |
+
"scr_dir2_threshold_50": 0.014285811599402364,
|
285 |
+
"scr_dir1_threshold_100": 0.33905571933328765,
|
286 |
+
"scr_metric_threshold_100": 0.33905571933328765,
|
287 |
+
"scr_dir2_threshold_100": 0.02380949677794379,
|
288 |
+
"scr_dir1_threshold_500": 0.1459226336960092,
|
289 |
+
"scr_metric_threshold_500": 0.1459226336960092,
|
290 |
+
"scr_dir2_threshold_500": -0.21904759742235502
|
291 |
+
}
|
292 |
+
],
|
293 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
294 |
+
"sae_lens_id": "custom_sae",
|
295 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4",
|
296 |
+
"sae_lens_version": "5.3.2",
|
297 |
+
"sae_cfg_dict": {
|
298 |
+
"model_name": "gemma-2-2b",
|
299 |
+
"d_in": 2304,
|
300 |
+
"d_sae": 16384,
|
301 |
+
"hook_layer": 12,
|
302 |
+
"hook_name": "blocks.12.hook_resid_post",
|
303 |
+
"context_size": null,
|
304 |
+
"hook_head_index": null,
|
305 |
+
"architecture": "matryoshka_batch_topk",
|
306 |
+
"apply_b_dec_to_input": null,
|
307 |
+
"finetuning_scaling_factor": null,
|
308 |
+
"activation_fn_str": "",
|
309 |
+
"prepend_bos": true,
|
310 |
+
"normalize_activations": "none",
|
311 |
+
"dtype": "bfloat16",
|
312 |
+
"device": "",
|
313 |
+
"dataset_path": "",
|
314 |
+
"dataset_trust_remote_code": true,
|
315 |
+
"seqpos_slice": [
|
316 |
+
null
|
317 |
+
],
|
318 |
+
"training_tokens": -100000,
|
319 |
+
"sae_lens_training_version": null,
|
320 |
+
"neuronpedia_id": null
|
321 |
+
},
|
322 |
+
"eval_result_unstructured": null
|
323 |
+
}
|
scr/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,323 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "scr",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"canrager/amazon_reviews_mcauley_1and5"
|
8 |
+
],
|
9 |
+
"perform_scr": true,
|
10 |
+
"early_stopping_patience": 20,
|
11 |
+
"train_set_size": 4000,
|
12 |
+
"test_set_size": 1000,
|
13 |
+
"context_length": 128,
|
14 |
+
"probe_train_batch_size": 16,
|
15 |
+
"probe_test_batch_size": 500,
|
16 |
+
"probe_epochs": 20,
|
17 |
+
"probe_lr": 0.001,
|
18 |
+
"probe_l1_penalty": 0.001,
|
19 |
+
"sae_batch_size": 125,
|
20 |
+
"llm_batch_size": 32,
|
21 |
+
"llm_dtype": "bfloat16",
|
22 |
+
"lower_vram_usage": false,
|
23 |
+
"model_name": "gemma-2-2b",
|
24 |
+
"n_values": [
|
25 |
+
2,
|
26 |
+
5,
|
27 |
+
10,
|
28 |
+
20,
|
29 |
+
50,
|
30 |
+
100,
|
31 |
+
500
|
32 |
+
],
|
33 |
+
"column1_vals_lookup": {
|
34 |
+
"LabHC/bias_in_bios_class_set1": [
|
35 |
+
[
|
36 |
+
"professor",
|
37 |
+
"nurse"
|
38 |
+
],
|
39 |
+
[
|
40 |
+
"architect",
|
41 |
+
"journalist"
|
42 |
+
],
|
43 |
+
[
|
44 |
+
"surgeon",
|
45 |
+
"psychologist"
|
46 |
+
],
|
47 |
+
[
|
48 |
+
"attorney",
|
49 |
+
"teacher"
|
50 |
+
]
|
51 |
+
],
|
52 |
+
"canrager/amazon_reviews_mcauley_1and5": [
|
53 |
+
[
|
54 |
+
"Books",
|
55 |
+
"CDs_and_Vinyl"
|
56 |
+
],
|
57 |
+
[
|
58 |
+
"Software",
|
59 |
+
"Electronics"
|
60 |
+
],
|
61 |
+
[
|
62 |
+
"Pet_Supplies",
|
63 |
+
"Office_Products"
|
64 |
+
],
|
65 |
+
[
|
66 |
+
"Industrial_and_Scientific",
|
67 |
+
"Toys_and_Games"
|
68 |
+
]
|
69 |
+
]
|
70 |
+
}
|
71 |
+
},
|
72 |
+
"eval_id": "418dfa2c-db68-408e-a396-a4493c37bac8",
|
73 |
+
"datetime_epoch_millis": 1737821138091,
|
74 |
+
"eval_result_metrics": {
|
75 |
+
"scr_metrics": {
|
76 |
+
"scr_dir1_threshold_2": 0.18868482887108878,
|
77 |
+
"scr_metric_threshold_2": 0.10145151794764225,
|
78 |
+
"scr_dir2_threshold_2": 0.09857749823078912,
|
79 |
+
"scr_dir1_threshold_5": 0.14830862633541209,
|
80 |
+
"scr_metric_threshold_5": 0.20069759422272948,
|
81 |
+
"scr_dir2_threshold_5": 0.19448718713184057,
|
82 |
+
"scr_dir1_threshold_10": 0.22555900222404784,
|
83 |
+
"scr_metric_threshold_10": 0.2552229192831546,
|
84 |
+
"scr_dir2_threshold_10": 0.2623581326463064,
|
85 |
+
"scr_dir1_threshold_20": 0.32814352899563703,
|
86 |
+
"scr_metric_threshold_20": 0.34590983810642967,
|
87 |
+
"scr_dir2_threshold_20": 0.32943219701586807,
|
88 |
+
"scr_dir1_threshold_50": 0.4077814415704477,
|
89 |
+
"scr_metric_threshold_50": 0.27142946819947916,
|
90 |
+
"scr_dir2_threshold_50": 0.22934128996584738,
|
91 |
+
"scr_dir1_threshold_100": -0.3702425280275639,
|
92 |
+
"scr_metric_threshold_100": 0.246033778932422,
|
93 |
+
"scr_dir2_threshold_100": 0.19674397374187466,
|
94 |
+
"scr_dir1_threshold_500": -0.7502976266384935,
|
95 |
+
"scr_metric_threshold_500": 0.042385799972817724,
|
96 |
+
"scr_dir2_threshold_500": -0.0542139336543357
|
97 |
+
}
|
98 |
+
},
|
99 |
+
"eval_result_details": [
|
100 |
+
{
|
101 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_professor_nurse_results",
|
102 |
+
"scr_dir1_threshold_2": 0.43749988358469727,
|
103 |
+
"scr_metric_threshold_2": 0.0566502484750167,
|
104 |
+
"scr_dir2_threshold_2": 0.0566502484750167,
|
105 |
+
"scr_dir1_threshold_5": 0.3437501746229541,
|
106 |
+
"scr_metric_threshold_5": 0.10837430540772658,
|
107 |
+
"scr_dir2_threshold_5": 0.10837430540772658,
|
108 |
+
"scr_dir1_threshold_10": 0.578124912688523,
|
109 |
+
"scr_metric_threshold_10": 0.07389155184943222,
|
110 |
+
"scr_dir2_threshold_10": 0.07389155184943222,
|
111 |
+
"scr_dir1_threshold_20": 0.2656252619344312,
|
112 |
+
"scr_metric_threshold_20": 0.2438422972744823,
|
113 |
+
"scr_dir2_threshold_20": 0.2438422972744823,
|
114 |
+
"scr_dir1_threshold_50": 0.5,
|
115 |
+
"scr_metric_threshold_50": 0.23891625254163884,
|
116 |
+
"scr_dir2_threshold_50": 0.23891625254163884,
|
117 |
+
"scr_dir1_threshold_100": -0.6249993015081835,
|
118 |
+
"scr_metric_threshold_100": 0.24137927490806058,
|
119 |
+
"scr_dir2_threshold_100": 0.24137927490806058,
|
120 |
+
"scr_dir1_threshold_500": 0.43749988358469727,
|
121 |
+
"scr_metric_threshold_500": 0.37931028914123804,
|
122 |
+
"scr_dir2_threshold_500": 0.37931028914123804
|
123 |
+
},
|
124 |
+
{
|
125 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_architect_journalist_results",
|
126 |
+
"scr_dir1_threshold_2": 0.16822437195083229,
|
127 |
+
"scr_metric_threshold_2": 0.1239194296368033,
|
128 |
+
"scr_dir2_threshold_2": 0.1239194296368033,
|
129 |
+
"scr_dir1_threshold_5": -0.018691906357197845,
|
130 |
+
"scr_metric_threshold_5": 0.24207490122290864,
|
131 |
+
"scr_dir2_threshold_5": 0.24207490122290864,
|
132 |
+
"scr_dir1_threshold_10": -0.8224299533969636,
|
133 |
+
"scr_metric_threshold_10": 0.2651297027978454,
|
134 |
+
"scr_dir2_threshold_10": 0.2651297027978454,
|
135 |
+
"scr_dir1_threshold_20": 0.2710279072306566,
|
136 |
+
"scr_metric_threshold_20": 0.3141211132017588,
|
137 |
+
"scr_dir2_threshold_20": 0.3141211132017588,
|
138 |
+
"scr_dir1_threshold_50": 0.6915888371077373,
|
139 |
+
"scr_metric_threshold_50": -0.2997117333889414,
|
140 |
+
"scr_dir2_threshold_50": -0.2997117333889414,
|
141 |
+
"scr_dir1_threshold_100": -2.579439627175709,
|
142 |
+
"scr_metric_threshold_100": -0.3227665349638782,
|
143 |
+
"scr_dir2_threshold_100": -0.3227665349638782,
|
144 |
+
"scr_dir1_threshold_500": -3.252337114979038,
|
145 |
+
"scr_metric_threshold_500": -0.38616715340929975,
|
146 |
+
"scr_dir2_threshold_500": -0.38616715340929975
|
147 |
+
},
|
148 |
+
{
|
149 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_surgeon_psychologist_results",
|
150 |
+
"scr_dir1_threshold_2": 0.3749997671693945,
|
151 |
+
"scr_metric_threshold_2": 0.05316462152337083,
|
152 |
+
"scr_dir2_threshold_2": 0.05316462152337083,
|
153 |
+
"scr_dir1_threshold_5": -0.18749941792348626,
|
154 |
+
"scr_metric_threshold_5": 0.1772153044793565,
|
155 |
+
"scr_dir2_threshold_5": 0.1772153044793565,
|
156 |
+
"scr_dir1_threshold_10": 0.5625001164153027,
|
157 |
+
"scr_metric_threshold_10": 0.14936711191080249,
|
158 |
+
"scr_dir2_threshold_10": 0.14936711191080249,
|
159 |
+
"scr_dir1_threshold_20": 0.5,
|
160 |
+
"scr_metric_threshold_20": 0.1873419062408471,
|
161 |
+
"scr_dir2_threshold_20": 0.1873419062408471,
|
162 |
+
"scr_dir1_threshold_50": -0.31249965075409175,
|
163 |
+
"scr_metric_threshold_50": 0.14936711191080249,
|
164 |
+
"scr_dir2_threshold_50": 0.14936711191080249,
|
165 |
+
"scr_dir1_threshold_100": -1.5312495925464404,
|
166 |
+
"scr_metric_threshold_100": 0.058227846955206435,
|
167 |
+
"scr_dir2_threshold_100": 0.058227846955206435,
|
168 |
+
"scr_dir1_threshold_500": -3.7968739231584494,
|
169 |
+
"scr_metric_threshold_500": -0.12405053205816628,
|
170 |
+
"scr_dir2_threshold_500": -0.12405053205816628
|
171 |
+
},
|
172 |
+
{
|
173 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_scr_attorney_teacher_results",
|
174 |
+
"scr_dir1_threshold_2": 0.12598447369816684,
|
175 |
+
"scr_metric_threshold_2": 0.04747783927687718,
|
176 |
+
"scr_dir2_threshold_2": 0.04747783927687718,
|
177 |
+
"scr_dir1_threshold_5": 0.14173229824654446,
|
178 |
+
"scr_metric_threshold_5": 0.1127597798483955,
|
179 |
+
"scr_dir2_threshold_5": 0.1127597798483955,
|
180 |
+
"scr_dir1_threshold_10": 0.7480315219314528,
|
181 |
+
"scr_metric_threshold_10": 0.18100904115762476,
|
182 |
+
"scr_dir2_threshold_10": 0.18100904115762476,
|
183 |
+
"scr_dir1_threshold_20": 0.5590552807119891,
|
184 |
+
"scr_metric_threshold_20": 0.19584582171455495,
|
185 |
+
"scr_dir2_threshold_20": 0.19584582171455495,
|
186 |
+
"scr_dir1_threshold_50": 0.36220465789054995,
|
187 |
+
"scr_metric_threshold_50": -0.3323441674646378,
|
188 |
+
"scr_dir2_threshold_50": -0.3323441674646378,
|
189 |
+
"scr_dir1_threshold_100": 0.259842390342736,
|
190 |
+
"scr_metric_threshold_100": 0.1127597798483955,
|
191 |
+
"scr_dir2_threshold_100": 0.1127597798483955,
|
192 |
+
"scr_dir1_threshold_500": -0.18897624121946377,
|
193 |
+
"scr_metric_threshold_500": 0.04747783927687718,
|
194 |
+
"scr_dir2_threshold_500": 0.04747783927687718
|
195 |
+
},
|
196 |
+
{
|
197 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Books_CDs_and_Vinyl_results",
|
198 |
+
"scr_dir1_threshold_2": 0.0,
|
199 |
+
"scr_metric_threshold_2": 0.20784303642428223,
|
200 |
+
"scr_dir2_threshold_2": 0.20784303642428223,
|
201 |
+
"scr_dir1_threshold_5": 0.20652165462480174,
|
202 |
+
"scr_metric_threshold_5": 0.21960782480441496,
|
203 |
+
"scr_dir2_threshold_5": 0.21960782480441496,
|
204 |
+
"scr_dir1_threshold_10": 0.19565223025013218,
|
205 |
+
"scr_metric_threshold_10": 0.4823528174298009,
|
206 |
+
"scr_dir2_threshold_10": 0.4823528174298009,
|
207 |
+
"scr_dir1_threshold_20": 0.17391305756253303,
|
208 |
+
"scr_metric_threshold_20": 0.5803921751955851,
|
209 |
+
"scr_dir2_threshold_20": 0.5803921751955851,
|
210 |
+
"scr_dir1_threshold_50": 0.6630433092496035,
|
211 |
+
"scr_metric_threshold_50": 0.780392175195585,
|
212 |
+
"scr_dir2_threshold_50": 0.780392175195585,
|
213 |
+
"scr_dir1_threshold_100": 0.42391305756253306,
|
214 |
+
"scr_metric_threshold_100": 0.7882352116198673,
|
215 |
+
"scr_dir2_threshold_100": 0.7882352116198673,
|
216 |
+
"scr_dir1_threshold_500": 0.021739172687599125,
|
217 |
+
"scr_metric_threshold_500": -0.6980393577657841,
|
218 |
+
"scr_dir2_threshold_500": -0.6980393577657841
|
219 |
+
},
|
220 |
+
{
|
221 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Software_Electronics_results",
|
222 |
+
"scr_dir1_threshold_2": 0.08762865692919337,
|
223 |
+
"scr_metric_threshold_2": 0.07258066454365199,
|
224 |
+
"scr_dir2_threshold_2": 0.07258066454365199,
|
225 |
+
"scr_dir1_threshold_5": 0.2061853642949754,
|
226 |
+
"scr_metric_threshold_5": 0.18951607282340924,
|
227 |
+
"scr_dir2_threshold_5": 0.18951607282340924,
|
228 |
+
"scr_dir1_threshold_10": 0.24742256005014163,
|
229 |
+
"scr_metric_threshold_10": 0.2782258006369044,
|
230 |
+
"scr_dir2_threshold_10": 0.2782258006369044,
|
231 |
+
"scr_dir1_threshold_20": 0.3195874990014686,
|
232 |
+
"scr_metric_threshold_20": 0.32661299044643394,
|
233 |
+
"scr_dir2_threshold_20": 0.32661299044643394,
|
234 |
+
"scr_dir1_threshold_50": 0.47938140212241687,
|
235 |
+
"scr_metric_threshold_50": 0.6209678543531815,
|
236 |
+
"scr_dir2_threshold_50": 0.6209678543531815,
|
237 |
+
"scr_dir1_threshold_100": 0.13917499800293723,
|
238 |
+
"scr_metric_threshold_100": 0.1653225980892868,
|
239 |
+
"scr_dir2_threshold_100": 0.1653225980892868,
|
240 |
+
"scr_dir1_threshold_500": 0.12886585268435963,
|
241 |
+
"scr_metric_threshold_500": 0.5887097278134952,
|
242 |
+
"scr_dir2_threshold_500": 0.5887097278134952
|
243 |
+
},
|
244 |
+
{
|
245 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Pet_Supplies_Office_Products_results",
|
246 |
+
"scr_dir1_threshold_2": 0.22072082956776967,
|
247 |
+
"scr_metric_threshold_2": 0.1555556556324794,
|
248 |
+
"scr_dir2_threshold_2": 0.1555556556324794,
|
249 |
+
"scr_dir1_threshold_5": 0.38288278129230385,
|
250 |
+
"scr_metric_threshold_5": 0.44444450331322316,
|
251 |
+
"scr_dir2_threshold_5": 0.44444450331322316,
|
252 |
+
"scr_dir1_threshold_10": 0.25225222322637253,
|
253 |
+
"scr_metric_threshold_10": 0.568888921855405,
|
254 |
+
"scr_dir2_threshold_10": 0.568888921855405,
|
255 |
+
"scr_dir1_threshold_20": 0.26126111613186265,
|
256 |
+
"scr_metric_threshold_20": 0.6444442913856198,
|
257 |
+
"scr_dir2_threshold_20": 0.6444442913856198,
|
258 |
+
"scr_dir1_threshold_50": 0.5180180543003676,
|
259 |
+
"scr_metric_threshold_50": 0.6533333298012066,
|
260 |
+
"scr_dir2_threshold_50": 0.6533333298012066,
|
261 |
+
"scr_dir1_threshold_100": 0.5945947179545834,
|
262 |
+
"scr_metric_threshold_100": 0.568888921855405,
|
263 |
+
"scr_dir2_threshold_100": 0.568888921855405,
|
264 |
+
"scr_dir1_threshold_500": 0.35585583408644605,
|
265 |
+
"scr_metric_threshold_500": 0.24000006357828102,
|
266 |
+
"scr_dir2_threshold_500": 0.24000006357828102
|
267 |
+
},
|
268 |
+
{
|
269 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_scr_Industrial_and_Scientific_Toys_and_Games_results",
|
270 |
+
"scr_dir1_threshold_2": 0.09442064806865635,
|
271 |
+
"scr_metric_threshold_2": 0.09442064806865635,
|
272 |
+
"scr_dir2_threshold_2": 0.07142849033383136,
|
273 |
+
"scr_dir1_threshold_5": 0.11158806188240133,
|
274 |
+
"scr_metric_threshold_5": 0.11158806188240133,
|
275 |
+
"scr_dir2_threshold_5": 0.06190480515528994,
|
276 |
+
"scr_dir1_threshold_10": 0.042918406627421406,
|
277 |
+
"scr_metric_threshold_10": 0.042918406627421406,
|
278 |
+
"scr_dir2_threshold_10": 0.10000011353263609,
|
279 |
+
"scr_dir1_threshold_20": 0.27467810939215553,
|
280 |
+
"scr_metric_threshold_20": 0.27467810939215553,
|
281 |
+
"scr_dir2_threshold_20": 0.14285698066766273,
|
282 |
+
"scr_dir1_threshold_50": 0.3605149226469983,
|
283 |
+
"scr_metric_threshold_50": 0.3605149226469983,
|
284 |
+
"scr_dir2_threshold_50": 0.02380949677794379,
|
285 |
+
"scr_dir1_threshold_100": 0.35622313314703263,
|
286 |
+
"scr_metric_threshold_100": 0.35622313314703263,
|
287 |
+
"scr_dir2_threshold_100": -0.03809530837734615,
|
288 |
+
"scr_dir1_threshold_500": 0.2918455232059005,
|
289 |
+
"scr_metric_threshold_500": 0.2918455232059005,
|
290 |
+
"scr_dir2_threshold_500": -0.4809523458113269
|
291 |
+
}
|
292 |
+
],
|
293 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
294 |
+
"sae_lens_id": "custom_sae",
|
295 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5",
|
296 |
+
"sae_lens_version": "5.3.2",
|
297 |
+
"sae_cfg_dict": {
|
298 |
+
"model_name": "gemma-2-2b",
|
299 |
+
"d_in": 2304,
|
300 |
+
"d_sae": 16384,
|
301 |
+
"hook_layer": 12,
|
302 |
+
"hook_name": "blocks.12.hook_resid_post",
|
303 |
+
"context_size": null,
|
304 |
+
"hook_head_index": null,
|
305 |
+
"architecture": "matryoshka_batch_topk",
|
306 |
+
"apply_b_dec_to_input": null,
|
307 |
+
"finetuning_scaling_factor": null,
|
308 |
+
"activation_fn_str": "",
|
309 |
+
"prepend_bos": true,
|
310 |
+
"normalize_activations": "none",
|
311 |
+
"dtype": "bfloat16",
|
312 |
+
"device": "",
|
313 |
+
"dataset_path": "",
|
314 |
+
"dataset_trust_remote_code": true,
|
315 |
+
"seqpos_slice": [
|
316 |
+
null
|
317 |
+
],
|
318 |
+
"training_tokens": -100000,
|
319 |
+
"sae_lens_training_version": null,
|
320 |
+
"neuronpedia_id": null
|
321 |
+
},
|
322 |
+
"eval_result_unstructured": null
|
323 |
+
}
|
sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,670 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "sparse_probing",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"LabHC/bias_in_bios_class_set2",
|
8 |
+
"LabHC/bias_in_bios_class_set3",
|
9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
11 |
+
"codeparrot/github-code",
|
12 |
+
"fancyzhx/ag_news",
|
13 |
+
"Helsinki-NLP/europarl"
|
14 |
+
],
|
15 |
+
"probe_train_set_size": 4000,
|
16 |
+
"probe_test_set_size": 1000,
|
17 |
+
"context_length": 128,
|
18 |
+
"sae_batch_size": 125,
|
19 |
+
"llm_batch_size": 32,
|
20 |
+
"llm_dtype": "bfloat16",
|
21 |
+
"model_name": "gemma-2-2b",
|
22 |
+
"k_values": [
|
23 |
+
1,
|
24 |
+
2,
|
25 |
+
5
|
26 |
+
],
|
27 |
+
"lower_vram_usage": false
|
28 |
+
},
|
29 |
+
"eval_id": "bb0bfb83-0672-4b53-b1c5-18eb0127bf65",
|
30 |
+
"datetime_epoch_millis": 1737822636179,
|
31 |
+
"eval_result_metrics": {
|
32 |
+
"llm": {
|
33 |
+
"llm_test_accuracy": 0.9590937938541175,
|
34 |
+
"llm_top_1_test_accuracy": 0.66480625,
|
35 |
+
"llm_top_2_test_accuracy": 0.7173375000000001,
|
36 |
+
"llm_top_5_test_accuracy": 0.78036875,
|
37 |
+
"llm_top_10_test_accuracy": null,
|
38 |
+
"llm_top_20_test_accuracy": null,
|
39 |
+
"llm_top_50_test_accuracy": null,
|
40 |
+
"llm_top_100_test_accuracy": null
|
41 |
+
},
|
42 |
+
"sae": {
|
43 |
+
"sae_test_accuracy": 0.9489687934517861,
|
44 |
+
"sae_top_1_test_accuracy": 0.76145,
|
45 |
+
"sae_top_2_test_accuracy": 0.8044250000000001,
|
46 |
+
"sae_top_5_test_accuracy": 0.85015625,
|
47 |
+
"sae_top_10_test_accuracy": null,
|
48 |
+
"sae_top_20_test_accuracy": null,
|
49 |
+
"sae_top_50_test_accuracy": null,
|
50 |
+
"sae_top_100_test_accuracy": null
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"eval_result_details": [
|
54 |
+
{
|
55 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
56 |
+
"llm_test_accuracy": 0.9694000363349915,
|
57 |
+
"llm_top_1_test_accuracy": 0.6436000000000001,
|
58 |
+
"llm_top_2_test_accuracy": 0.6874,
|
59 |
+
"llm_top_5_test_accuracy": 0.7908,
|
60 |
+
"llm_top_10_test_accuracy": null,
|
61 |
+
"llm_top_20_test_accuracy": null,
|
62 |
+
"llm_top_50_test_accuracy": null,
|
63 |
+
"llm_top_100_test_accuracy": null,
|
64 |
+
"sae_test_accuracy": 0.9582000374794006,
|
65 |
+
"sae_top_1_test_accuracy": 0.7978,
|
66 |
+
"sae_top_2_test_accuracy": 0.8028000000000001,
|
67 |
+
"sae_top_5_test_accuracy": 0.8460000000000001,
|
68 |
+
"sae_top_10_test_accuracy": null,
|
69 |
+
"sae_top_20_test_accuracy": null,
|
70 |
+
"sae_top_50_test_accuracy": null,
|
71 |
+
"sae_top_100_test_accuracy": null
|
72 |
+
},
|
73 |
+
{
|
74 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
75 |
+
"llm_test_accuracy": 0.9544000387191772,
|
76 |
+
"llm_top_1_test_accuracy": 0.6764,
|
77 |
+
"llm_top_2_test_accuracy": 0.7150000000000001,
|
78 |
+
"llm_top_5_test_accuracy": 0.7592000000000001,
|
79 |
+
"llm_top_10_test_accuracy": null,
|
80 |
+
"llm_top_20_test_accuracy": null,
|
81 |
+
"llm_top_50_test_accuracy": null,
|
82 |
+
"llm_top_100_test_accuracy": null,
|
83 |
+
"sae_test_accuracy": 0.9412000417709351,
|
84 |
+
"sae_top_1_test_accuracy": 0.767,
|
85 |
+
"sae_top_2_test_accuracy": 0.7703999999999999,
|
86 |
+
"sae_top_5_test_accuracy": 0.8230000000000001,
|
87 |
+
"sae_top_10_test_accuracy": null,
|
88 |
+
"sae_top_20_test_accuracy": null,
|
89 |
+
"sae_top_50_test_accuracy": null,
|
90 |
+
"sae_top_100_test_accuracy": null
|
91 |
+
},
|
92 |
+
{
|
93 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
94 |
+
"llm_test_accuracy": 0.9332000494003296,
|
95 |
+
"llm_top_1_test_accuracy": 0.6884,
|
96 |
+
"llm_top_2_test_accuracy": 0.737,
|
97 |
+
"llm_top_5_test_accuracy": 0.7492,
|
98 |
+
"llm_top_10_test_accuracy": null,
|
99 |
+
"llm_top_20_test_accuracy": null,
|
100 |
+
"llm_top_50_test_accuracy": null,
|
101 |
+
"llm_top_100_test_accuracy": null,
|
102 |
+
"sae_test_accuracy": 0.9212000370025635,
|
103 |
+
"sae_top_1_test_accuracy": 0.7862,
|
104 |
+
"sae_top_2_test_accuracy": 0.8128,
|
105 |
+
"sae_top_5_test_accuracy": 0.8482000000000001,
|
106 |
+
"sae_top_10_test_accuracy": null,
|
107 |
+
"sae_top_20_test_accuracy": null,
|
108 |
+
"sae_top_50_test_accuracy": null,
|
109 |
+
"sae_top_100_test_accuracy": null
|
110 |
+
},
|
111 |
+
{
|
112 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
113 |
+
"llm_test_accuracy": 0.9164000511169433,
|
114 |
+
"llm_top_1_test_accuracy": 0.598,
|
115 |
+
"llm_top_2_test_accuracy": 0.6406,
|
116 |
+
"llm_top_5_test_accuracy": 0.6748000000000001,
|
117 |
+
"llm_top_10_test_accuracy": null,
|
118 |
+
"llm_top_20_test_accuracy": null,
|
119 |
+
"llm_top_50_test_accuracy": null,
|
120 |
+
"llm_top_100_test_accuracy": null,
|
121 |
+
"sae_test_accuracy": 0.9056000471115112,
|
122 |
+
"sae_top_1_test_accuracy": 0.6504000000000001,
|
123 |
+
"sae_top_2_test_accuracy": 0.7308,
|
124 |
+
"sae_top_5_test_accuracy": 0.7818,
|
125 |
+
"sae_top_10_test_accuracy": null,
|
126 |
+
"sae_top_20_test_accuracy": null,
|
127 |
+
"sae_top_50_test_accuracy": null,
|
128 |
+
"sae_top_100_test_accuracy": null
|
129 |
+
},
|
130 |
+
{
|
131 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
132 |
+
"llm_test_accuracy": 0.9820000529289246,
|
133 |
+
"llm_top_1_test_accuracy": 0.672,
|
134 |
+
"llm_top_2_test_accuracy": 0.724,
|
135 |
+
"llm_top_5_test_accuracy": 0.766,
|
136 |
+
"llm_top_10_test_accuracy": null,
|
137 |
+
"llm_top_20_test_accuracy": null,
|
138 |
+
"llm_top_50_test_accuracy": null,
|
139 |
+
"llm_top_100_test_accuracy": null,
|
140 |
+
"sae_test_accuracy": 0.9585000574588776,
|
141 |
+
"sae_top_1_test_accuracy": 0.791,
|
142 |
+
"sae_top_2_test_accuracy": 0.846,
|
143 |
+
"sae_top_5_test_accuracy": 0.895,
|
144 |
+
"sae_top_10_test_accuracy": null,
|
145 |
+
"sae_top_20_test_accuracy": null,
|
146 |
+
"sae_top_50_test_accuracy": null,
|
147 |
+
"sae_top_100_test_accuracy": null
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"dataset_name": "codeparrot/github-code_results",
|
151 |
+
"llm_test_accuracy": 0.9672000408172607,
|
152 |
+
"llm_top_1_test_accuracy": 0.6668000000000001,
|
153 |
+
"llm_top_2_test_accuracy": 0.6896,
|
154 |
+
"llm_top_5_test_accuracy": 0.7656000000000001,
|
155 |
+
"llm_top_10_test_accuracy": null,
|
156 |
+
"llm_top_20_test_accuracy": null,
|
157 |
+
"llm_top_50_test_accuracy": null,
|
158 |
+
"llm_top_100_test_accuracy": null,
|
159 |
+
"sae_test_accuracy": 0.9656000375747681,
|
160 |
+
"sae_top_1_test_accuracy": 0.6324,
|
161 |
+
"sae_top_2_test_accuracy": 0.6910000000000001,
|
162 |
+
"sae_top_5_test_accuracy": 0.7904,
|
163 |
+
"sae_top_10_test_accuracy": null,
|
164 |
+
"sae_top_20_test_accuracy": null,
|
165 |
+
"sae_top_50_test_accuracy": null,
|
166 |
+
"sae_top_100_test_accuracy": null
|
167 |
+
},
|
168 |
+
{
|
169 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
170 |
+
"llm_test_accuracy": 0.9507500380277634,
|
171 |
+
"llm_top_1_test_accuracy": 0.63625,
|
172 |
+
"llm_top_2_test_accuracy": 0.7655000000000001,
|
173 |
+
"llm_top_5_test_accuracy": 0.82475,
|
174 |
+
"llm_top_10_test_accuracy": null,
|
175 |
+
"llm_top_20_test_accuracy": null,
|
176 |
+
"llm_top_50_test_accuracy": null,
|
177 |
+
"llm_top_100_test_accuracy": null,
|
178 |
+
"sae_test_accuracy": 0.9472500383853912,
|
179 |
+
"sae_top_1_test_accuracy": 0.7649999999999999,
|
180 |
+
"sae_top_2_test_accuracy": 0.853,
|
181 |
+
"sae_top_5_test_accuracy": 0.8652499999999999,
|
182 |
+
"sae_top_10_test_accuracy": null,
|
183 |
+
"sae_top_20_test_accuracy": null,
|
184 |
+
"sae_top_50_test_accuracy": null,
|
185 |
+
"sae_top_100_test_accuracy": null
|
186 |
+
},
|
187 |
+
{
|
188 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
189 |
+
"llm_test_accuracy": 0.9994000434875489,
|
190 |
+
"llm_top_1_test_accuracy": 0.737,
|
191 |
+
"llm_top_2_test_accuracy": 0.7796,
|
192 |
+
"llm_top_5_test_accuracy": 0.9126,
|
193 |
+
"llm_top_10_test_accuracy": null,
|
194 |
+
"llm_top_20_test_accuracy": null,
|
195 |
+
"llm_top_50_test_accuracy": null,
|
196 |
+
"llm_top_100_test_accuracy": null,
|
197 |
+
"sae_test_accuracy": 0.9942000508308411,
|
198 |
+
"sae_top_1_test_accuracy": 0.9018,
|
199 |
+
"sae_top_2_test_accuracy": 0.9286,
|
200 |
+
"sae_top_5_test_accuracy": 0.9516,
|
201 |
+
"sae_top_10_test_accuracy": null,
|
202 |
+
"sae_top_20_test_accuracy": null,
|
203 |
+
"sae_top_50_test_accuracy": null,
|
204 |
+
"sae_top_100_test_accuracy": null
|
205 |
+
}
|
206 |
+
],
|
207 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
208 |
+
"sae_lens_id": "custom_sae",
|
209 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0",
|
210 |
+
"sae_lens_version": "5.3.2",
|
211 |
+
"sae_cfg_dict": {
|
212 |
+
"model_name": "gemma-2-2b",
|
213 |
+
"d_in": 2304,
|
214 |
+
"d_sae": 16384,
|
215 |
+
"hook_layer": 12,
|
216 |
+
"hook_name": "blocks.12.hook_resid_post",
|
217 |
+
"context_size": null,
|
218 |
+
"hook_head_index": null,
|
219 |
+
"architecture": "matryoshka_batch_topk",
|
220 |
+
"apply_b_dec_to_input": null,
|
221 |
+
"finetuning_scaling_factor": null,
|
222 |
+
"activation_fn_str": "",
|
223 |
+
"prepend_bos": true,
|
224 |
+
"normalize_activations": "none",
|
225 |
+
"dtype": "bfloat16",
|
226 |
+
"device": "",
|
227 |
+
"dataset_path": "",
|
228 |
+
"dataset_trust_remote_code": true,
|
229 |
+
"seqpos_slice": [
|
230 |
+
null
|
231 |
+
],
|
232 |
+
"training_tokens": -100000,
|
233 |
+
"sae_lens_training_version": null,
|
234 |
+
"neuronpedia_id": null
|
235 |
+
},
|
236 |
+
"eval_result_unstructured": {
|
237 |
+
"LabHC/bias_in_bios_class_set1_results": {
|
238 |
+
"sae_test_accuracy": {
|
239 |
+
"0": 0.9440000653266907,
|
240 |
+
"1": 0.9570000171661377,
|
241 |
+
"2": 0.9410000443458557,
|
242 |
+
"6": 0.9830000400543213,
|
243 |
+
"9": 0.9660000205039978
|
244 |
+
},
|
245 |
+
"llm_test_accuracy": {
|
246 |
+
"0": 0.9510000348091125,
|
247 |
+
"1": 0.9670000672340393,
|
248 |
+
"2": 0.9520000219345093,
|
249 |
+
"6": 0.9930000305175781,
|
250 |
+
"9": 0.984000027179718
|
251 |
+
},
|
252 |
+
"llm_top_1_test_accuracy": {
|
253 |
+
"0": 0.568,
|
254 |
+
"1": 0.629,
|
255 |
+
"2": 0.679,
|
256 |
+
"6": 0.791,
|
257 |
+
"9": 0.551
|
258 |
+
},
|
259 |
+
"llm_top_2_test_accuracy": {
|
260 |
+
"0": 0.585,
|
261 |
+
"1": 0.666,
|
262 |
+
"2": 0.673,
|
263 |
+
"6": 0.801,
|
264 |
+
"9": 0.712
|
265 |
+
},
|
266 |
+
"llm_top_5_test_accuracy": {
|
267 |
+
"0": 0.72,
|
268 |
+
"1": 0.707,
|
269 |
+
"2": 0.764,
|
270 |
+
"6": 0.899,
|
271 |
+
"9": 0.864
|
272 |
+
},
|
273 |
+
"sae_top_1_test_accuracy": {
|
274 |
+
"0": 0.6,
|
275 |
+
"1": 0.624,
|
276 |
+
"2": 0.865,
|
277 |
+
"6": 0.979,
|
278 |
+
"9": 0.921
|
279 |
+
},
|
280 |
+
"sae_top_2_test_accuracy": {
|
281 |
+
"0": 0.627,
|
282 |
+
"1": 0.63,
|
283 |
+
"2": 0.865,
|
284 |
+
"6": 0.975,
|
285 |
+
"9": 0.917
|
286 |
+
},
|
287 |
+
"sae_top_5_test_accuracy": {
|
288 |
+
"0": 0.693,
|
289 |
+
"1": 0.769,
|
290 |
+
"2": 0.871,
|
291 |
+
"6": 0.977,
|
292 |
+
"9": 0.92
|
293 |
+
}
|
294 |
+
},
|
295 |
+
"LabHC/bias_in_bios_class_set2_results": {
|
296 |
+
"sae_test_accuracy": {
|
297 |
+
"11": 0.9550000429153442,
|
298 |
+
"13": 0.9490000605583191,
|
299 |
+
"14": 0.9500000476837158,
|
300 |
+
"18": 0.906000018119812,
|
301 |
+
"19": 0.9460000395774841
|
302 |
+
},
|
303 |
+
"llm_test_accuracy": {
|
304 |
+
"11": 0.9650000333786011,
|
305 |
+
"13": 0.9520000219345093,
|
306 |
+
"14": 0.9550000429153442,
|
307 |
+
"18": 0.9360000491142273,
|
308 |
+
"19": 0.9640000462532043
|
309 |
+
},
|
310 |
+
"llm_top_1_test_accuracy": {
|
311 |
+
"11": 0.558,
|
312 |
+
"13": 0.673,
|
313 |
+
"14": 0.656,
|
314 |
+
"18": 0.702,
|
315 |
+
"19": 0.793
|
316 |
+
},
|
317 |
+
"llm_top_2_test_accuracy": {
|
318 |
+
"11": 0.686,
|
319 |
+
"13": 0.713,
|
320 |
+
"14": 0.687,
|
321 |
+
"18": 0.724,
|
322 |
+
"19": 0.765
|
323 |
+
},
|
324 |
+
"llm_top_5_test_accuracy": {
|
325 |
+
"11": 0.782,
|
326 |
+
"13": 0.742,
|
327 |
+
"14": 0.716,
|
328 |
+
"18": 0.725,
|
329 |
+
"19": 0.831
|
330 |
+
},
|
331 |
+
"sae_top_1_test_accuracy": {
|
332 |
+
"11": 0.738,
|
333 |
+
"13": 0.693,
|
334 |
+
"14": 0.871,
|
335 |
+
"18": 0.702,
|
336 |
+
"19": 0.831
|
337 |
+
},
|
338 |
+
"sae_top_2_test_accuracy": {
|
339 |
+
"11": 0.733,
|
340 |
+
"13": 0.689,
|
341 |
+
"14": 0.87,
|
342 |
+
"18": 0.727,
|
343 |
+
"19": 0.833
|
344 |
+
},
|
345 |
+
"sae_top_5_test_accuracy": {
|
346 |
+
"11": 0.843,
|
347 |
+
"13": 0.745,
|
348 |
+
"14": 0.865,
|
349 |
+
"18": 0.761,
|
350 |
+
"19": 0.901
|
351 |
+
}
|
352 |
+
},
|
353 |
+
"LabHC/bias_in_bios_class_set3_results": {
|
354 |
+
"sae_test_accuracy": {
|
355 |
+
"20": 0.9510000348091125,
|
356 |
+
"21": 0.9140000343322754,
|
357 |
+
"22": 0.9080000519752502,
|
358 |
+
"25": 0.9500000476837158,
|
359 |
+
"26": 0.8830000162124634
|
360 |
+
},
|
361 |
+
"llm_test_accuracy": {
|
362 |
+
"20": 0.9610000252723694,
|
363 |
+
"21": 0.9270000457763672,
|
364 |
+
"22": 0.921000063419342,
|
365 |
+
"25": 0.968000054359436,
|
366 |
+
"26": 0.8890000581741333
|
367 |
+
},
|
368 |
+
"llm_top_1_test_accuracy": {
|
369 |
+
"20": 0.713,
|
370 |
+
"21": 0.766,
|
371 |
+
"22": 0.639,
|
372 |
+
"25": 0.69,
|
373 |
+
"26": 0.634
|
374 |
+
},
|
375 |
+
"llm_top_2_test_accuracy": {
|
376 |
+
"20": 0.804,
|
377 |
+
"21": 0.777,
|
378 |
+
"22": 0.68,
|
379 |
+
"25": 0.759,
|
380 |
+
"26": 0.665
|
381 |
+
},
|
382 |
+
"llm_top_5_test_accuracy": {
|
383 |
+
"20": 0.813,
|
384 |
+
"21": 0.777,
|
385 |
+
"22": 0.686,
|
386 |
+
"25": 0.798,
|
387 |
+
"26": 0.672
|
388 |
+
},
|
389 |
+
"sae_top_1_test_accuracy": {
|
390 |
+
"20": 0.893,
|
391 |
+
"21": 0.755,
|
392 |
+
"22": 0.794,
|
393 |
+
"25": 0.874,
|
394 |
+
"26": 0.615
|
395 |
+
},
|
396 |
+
"sae_top_2_test_accuracy": {
|
397 |
+
"20": 0.891,
|
398 |
+
"21": 0.782,
|
399 |
+
"22": 0.818,
|
400 |
+
"25": 0.861,
|
401 |
+
"26": 0.712
|
402 |
+
},
|
403 |
+
"sae_top_5_test_accuracy": {
|
404 |
+
"20": 0.906,
|
405 |
+
"21": 0.815,
|
406 |
+
"22": 0.842,
|
407 |
+
"25": 0.893,
|
408 |
+
"26": 0.785
|
409 |
+
}
|
410 |
+
},
|
411 |
+
"canrager/amazon_reviews_mcauley_1and5_results": {
|
412 |
+
"sae_test_accuracy": {
|
413 |
+
"1": 0.9320000410079956,
|
414 |
+
"2": 0.9250000715255737,
|
415 |
+
"3": 0.9000000357627869,
|
416 |
+
"5": 0.9050000309944153,
|
417 |
+
"6": 0.8660000562667847
|
418 |
+
},
|
419 |
+
"llm_test_accuracy": {
|
420 |
+
"1": 0.9480000734329224,
|
421 |
+
"2": 0.9350000619888306,
|
422 |
+
"3": 0.9140000343322754,
|
423 |
+
"5": 0.9110000729560852,
|
424 |
+
"6": 0.8740000128746033
|
425 |
+
},
|
426 |
+
"llm_top_1_test_accuracy": {
|
427 |
+
"1": 0.668,
|
428 |
+
"2": 0.573,
|
429 |
+
"3": 0.591,
|
430 |
+
"5": 0.568,
|
431 |
+
"6": 0.59
|
432 |
+
},
|
433 |
+
"llm_top_2_test_accuracy": {
|
434 |
+
"1": 0.745,
|
435 |
+
"2": 0.636,
|
436 |
+
"3": 0.606,
|
437 |
+
"5": 0.595,
|
438 |
+
"6": 0.621
|
439 |
+
},
|
440 |
+
"llm_top_5_test_accuracy": {
|
441 |
+
"1": 0.778,
|
442 |
+
"2": 0.643,
|
443 |
+
"3": 0.608,
|
444 |
+
"5": 0.66,
|
445 |
+
"6": 0.685
|
446 |
+
},
|
447 |
+
"sae_top_1_test_accuracy": {
|
448 |
+
"1": 0.669,
|
449 |
+
"2": 0.764,
|
450 |
+
"3": 0.605,
|
451 |
+
"5": 0.558,
|
452 |
+
"6": 0.656
|
453 |
+
},
|
454 |
+
"sae_top_2_test_accuracy": {
|
455 |
+
"1": 0.788,
|
456 |
+
"2": 0.787,
|
457 |
+
"3": 0.597,
|
458 |
+
"5": 0.731,
|
459 |
+
"6": 0.751
|
460 |
+
},
|
461 |
+
"sae_top_5_test_accuracy": {
|
462 |
+
"1": 0.888,
|
463 |
+
"2": 0.795,
|
464 |
+
"3": 0.759,
|
465 |
+
"5": 0.713,
|
466 |
+
"6": 0.754
|
467 |
+
}
|
468 |
+
},
|
469 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment_results": {
|
470 |
+
"sae_test_accuracy": {
|
471 |
+
"1.0": 0.9590000510215759,
|
472 |
+
"5.0": 0.9580000638961792
|
473 |
+
},
|
474 |
+
"llm_test_accuracy": {
|
475 |
+
"1.0": 0.9810000658035278,
|
476 |
+
"5.0": 0.9830000400543213
|
477 |
+
},
|
478 |
+
"llm_top_1_test_accuracy": {
|
479 |
+
"1.0": 0.672,
|
480 |
+
"5.0": 0.672
|
481 |
+
},
|
482 |
+
"llm_top_2_test_accuracy": {
|
483 |
+
"1.0": 0.724,
|
484 |
+
"5.0": 0.724
|
485 |
+
},
|
486 |
+
"llm_top_5_test_accuracy": {
|
487 |
+
"1.0": 0.766,
|
488 |
+
"5.0": 0.766
|
489 |
+
},
|
490 |
+
"sae_top_1_test_accuracy": {
|
491 |
+
"1.0": 0.791,
|
492 |
+
"5.0": 0.791
|
493 |
+
},
|
494 |
+
"sae_top_2_test_accuracy": {
|
495 |
+
"1.0": 0.846,
|
496 |
+
"5.0": 0.846
|
497 |
+
},
|
498 |
+
"sae_top_5_test_accuracy": {
|
499 |
+
"1.0": 0.895,
|
500 |
+
"5.0": 0.895
|
501 |
+
}
|
502 |
+
},
|
503 |
+
"codeparrot/github-code_results": {
|
504 |
+
"sae_test_accuracy": {
|
505 |
+
"C": 0.9460000395774841,
|
506 |
+
"Python": 0.9780000448226929,
|
507 |
+
"HTML": 0.984000027179718,
|
508 |
+
"Java": 0.9650000333786011,
|
509 |
+
"PHP": 0.9550000429153442
|
510 |
+
},
|
511 |
+
"llm_test_accuracy": {
|
512 |
+
"C": 0.9510000348091125,
|
513 |
+
"Python": 0.987000048160553,
|
514 |
+
"HTML": 0.984000027179718,
|
515 |
+
"Java": 0.968000054359436,
|
516 |
+
"PHP": 0.9460000395774841
|
517 |
+
},
|
518 |
+
"llm_top_1_test_accuracy": {
|
519 |
+
"C": 0.655,
|
520 |
+
"Python": 0.661,
|
521 |
+
"HTML": 0.802,
|
522 |
+
"Java": 0.622,
|
523 |
+
"PHP": 0.594
|
524 |
+
},
|
525 |
+
"llm_top_2_test_accuracy": {
|
526 |
+
"C": 0.667,
|
527 |
+
"Python": 0.668,
|
528 |
+
"HTML": 0.798,
|
529 |
+
"Java": 0.674,
|
530 |
+
"PHP": 0.641
|
531 |
+
},
|
532 |
+
"llm_top_5_test_accuracy": {
|
533 |
+
"C": 0.78,
|
534 |
+
"Python": 0.73,
|
535 |
+
"HTML": 0.906,
|
536 |
+
"Java": 0.728,
|
537 |
+
"PHP": 0.684
|
538 |
+
},
|
539 |
+
"sae_top_1_test_accuracy": {
|
540 |
+
"C": 0.615,
|
541 |
+
"Python": 0.622,
|
542 |
+
"HTML": 0.694,
|
543 |
+
"Java": 0.625,
|
544 |
+
"PHP": 0.606
|
545 |
+
},
|
546 |
+
"sae_top_2_test_accuracy": {
|
547 |
+
"C": 0.624,
|
548 |
+
"Python": 0.653,
|
549 |
+
"HTML": 0.912,
|
550 |
+
"Java": 0.665,
|
551 |
+
"PHP": 0.601
|
552 |
+
},
|
553 |
+
"sae_top_5_test_accuracy": {
|
554 |
+
"C": 0.651,
|
555 |
+
"Python": 0.654,
|
556 |
+
"HTML": 0.944,
|
557 |
+
"Java": 0.771,
|
558 |
+
"PHP": 0.932
|
559 |
+
}
|
560 |
+
},
|
561 |
+
"fancyzhx/ag_news_results": {
|
562 |
+
"sae_test_accuracy": {
|
563 |
+
"0": 0.9410000443458557,
|
564 |
+
"1": 0.9790000319480896,
|
565 |
+
"2": 0.9220000505447388,
|
566 |
+
"3": 0.9470000267028809
|
567 |
+
},
|
568 |
+
"llm_test_accuracy": {
|
569 |
+
"0": 0.9360000491142273,
|
570 |
+
"1": 0.9890000224113464,
|
571 |
+
"2": 0.9280000329017639,
|
572 |
+
"3": 0.9500000476837158
|
573 |
+
},
|
574 |
+
"llm_top_1_test_accuracy": {
|
575 |
+
"0": 0.58,
|
576 |
+
"1": 0.676,
|
577 |
+
"2": 0.667,
|
578 |
+
"3": 0.622
|
579 |
+
},
|
580 |
+
"llm_top_2_test_accuracy": {
|
581 |
+
"0": 0.811,
|
582 |
+
"1": 0.798,
|
583 |
+
"2": 0.693,
|
584 |
+
"3": 0.76
|
585 |
+
},
|
586 |
+
"llm_top_5_test_accuracy": {
|
587 |
+
"0": 0.815,
|
588 |
+
"1": 0.884,
|
589 |
+
"2": 0.756,
|
590 |
+
"3": 0.844
|
591 |
+
},
|
592 |
+
"sae_top_1_test_accuracy": {
|
593 |
+
"0": 0.824,
|
594 |
+
"1": 0.897,
|
595 |
+
"2": 0.704,
|
596 |
+
"3": 0.635
|
597 |
+
},
|
598 |
+
"sae_top_2_test_accuracy": {
|
599 |
+
"0": 0.854,
|
600 |
+
"1": 0.946,
|
601 |
+
"2": 0.833,
|
602 |
+
"3": 0.779
|
603 |
+
},
|
604 |
+
"sae_top_5_test_accuracy": {
|
605 |
+
"0": 0.864,
|
606 |
+
"1": 0.948,
|
607 |
+
"2": 0.833,
|
608 |
+
"3": 0.816
|
609 |
+
}
|
610 |
+
},
|
611 |
+
"Helsinki-NLP/europarl_results": {
|
612 |
+
"sae_test_accuracy": {
|
613 |
+
"en": 0.9970000386238098,
|
614 |
+
"fr": 0.9950000643730164,
|
615 |
+
"de": 0.9920000433921814,
|
616 |
+
"es": 0.9960000514984131,
|
617 |
+
"nl": 0.9910000562667847
|
618 |
+
},
|
619 |
+
"llm_test_accuracy": {
|
620 |
+
"en": 1.0,
|
621 |
+
"fr": 0.999000072479248,
|
622 |
+
"de": 0.999000072479248,
|
623 |
+
"es": 1.0,
|
624 |
+
"nl": 0.999000072479248
|
625 |
+
},
|
626 |
+
"llm_top_1_test_accuracy": {
|
627 |
+
"en": 0.748,
|
628 |
+
"fr": 0.615,
|
629 |
+
"de": 0.761,
|
630 |
+
"es": 0.904,
|
631 |
+
"nl": 0.657
|
632 |
+
},
|
633 |
+
"llm_top_2_test_accuracy": {
|
634 |
+
"en": 0.829,
|
635 |
+
"fr": 0.583,
|
636 |
+
"de": 0.829,
|
637 |
+
"es": 0.914,
|
638 |
+
"nl": 0.743
|
639 |
+
},
|
640 |
+
"llm_top_5_test_accuracy": {
|
641 |
+
"en": 0.892,
|
642 |
+
"fr": 0.914,
|
643 |
+
"de": 0.917,
|
644 |
+
"es": 0.984,
|
645 |
+
"nl": 0.856
|
646 |
+
},
|
647 |
+
"sae_top_1_test_accuracy": {
|
648 |
+
"en": 0.999,
|
649 |
+
"fr": 0.979,
|
650 |
+
"de": 0.78,
|
651 |
+
"es": 0.987,
|
652 |
+
"nl": 0.764
|
653 |
+
},
|
654 |
+
"sae_top_2_test_accuracy": {
|
655 |
+
"en": 1.0,
|
656 |
+
"fr": 0.984,
|
657 |
+
"de": 0.902,
|
658 |
+
"es": 0.986,
|
659 |
+
"nl": 0.771
|
660 |
+
},
|
661 |
+
"sae_top_5_test_accuracy": {
|
662 |
+
"en": 0.999,
|
663 |
+
"fr": 0.991,
|
664 |
+
"de": 0.974,
|
665 |
+
"es": 0.986,
|
666 |
+
"nl": 0.808
|
667 |
+
}
|
668 |
+
}
|
669 |
+
}
|
670 |
+
}
|
sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,670 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "sparse_probing",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"LabHC/bias_in_bios_class_set2",
|
8 |
+
"LabHC/bias_in_bios_class_set3",
|
9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
11 |
+
"codeparrot/github-code",
|
12 |
+
"fancyzhx/ag_news",
|
13 |
+
"Helsinki-NLP/europarl"
|
14 |
+
],
|
15 |
+
"probe_train_set_size": 4000,
|
16 |
+
"probe_test_set_size": 1000,
|
17 |
+
"context_length": 128,
|
18 |
+
"sae_batch_size": 125,
|
19 |
+
"llm_batch_size": 32,
|
20 |
+
"llm_dtype": "bfloat16",
|
21 |
+
"model_name": "gemma-2-2b",
|
22 |
+
"k_values": [
|
23 |
+
1,
|
24 |
+
2,
|
25 |
+
5
|
26 |
+
],
|
27 |
+
"lower_vram_usage": false
|
28 |
+
},
|
29 |
+
"eval_id": "16d9d659-405c-4b74-a15b-93c546cf375a",
|
30 |
+
"datetime_epoch_millis": 1737822947171,
|
31 |
+
"eval_result_metrics": {
|
32 |
+
"llm": {
|
33 |
+
"llm_test_accuracy": 0.9590937938541175,
|
34 |
+
"llm_top_1_test_accuracy": 0.66480625,
|
35 |
+
"llm_top_2_test_accuracy": 0.7173375000000001,
|
36 |
+
"llm_top_5_test_accuracy": 0.78036875,
|
37 |
+
"llm_top_10_test_accuracy": null,
|
38 |
+
"llm_top_20_test_accuracy": null,
|
39 |
+
"llm_top_50_test_accuracy": null,
|
40 |
+
"llm_top_100_test_accuracy": null
|
41 |
+
},
|
42 |
+
"sae": {
|
43 |
+
"sae_test_accuracy": 0.9521000370383262,
|
44 |
+
"sae_top_1_test_accuracy": 0.77283125,
|
45 |
+
"sae_top_2_test_accuracy": 0.81136875,
|
46 |
+
"sae_top_5_test_accuracy": 0.87860625,
|
47 |
+
"sae_top_10_test_accuracy": null,
|
48 |
+
"sae_top_20_test_accuracy": null,
|
49 |
+
"sae_top_50_test_accuracy": null,
|
50 |
+
"sae_top_100_test_accuracy": null
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"eval_result_details": [
|
54 |
+
{
|
55 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
56 |
+
"llm_test_accuracy": 0.9694000363349915,
|
57 |
+
"llm_top_1_test_accuracy": 0.6436000000000001,
|
58 |
+
"llm_top_2_test_accuracy": 0.6874,
|
59 |
+
"llm_top_5_test_accuracy": 0.7908,
|
60 |
+
"llm_top_10_test_accuracy": null,
|
61 |
+
"llm_top_20_test_accuracy": null,
|
62 |
+
"llm_top_50_test_accuracy": null,
|
63 |
+
"llm_top_100_test_accuracy": null,
|
64 |
+
"sae_test_accuracy": 0.9600000381469727,
|
65 |
+
"sae_top_1_test_accuracy": 0.7566,
|
66 |
+
"sae_top_2_test_accuracy": 0.8055999999999999,
|
67 |
+
"sae_top_5_test_accuracy": 0.8630000000000001,
|
68 |
+
"sae_top_10_test_accuracy": null,
|
69 |
+
"sae_top_20_test_accuracy": null,
|
70 |
+
"sae_top_50_test_accuracy": null,
|
71 |
+
"sae_top_100_test_accuracy": null
|
72 |
+
},
|
73 |
+
{
|
74 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
75 |
+
"llm_test_accuracy": 0.9544000387191772,
|
76 |
+
"llm_top_1_test_accuracy": 0.6764,
|
77 |
+
"llm_top_2_test_accuracy": 0.7150000000000001,
|
78 |
+
"llm_top_5_test_accuracy": 0.7592000000000001,
|
79 |
+
"llm_top_10_test_accuracy": null,
|
80 |
+
"llm_top_20_test_accuracy": null,
|
81 |
+
"llm_top_50_test_accuracy": null,
|
82 |
+
"llm_top_100_test_accuracy": null,
|
83 |
+
"sae_test_accuracy": 0.9450000405311585,
|
84 |
+
"sae_top_1_test_accuracy": 0.7934,
|
85 |
+
"sae_top_2_test_accuracy": 0.7866000000000001,
|
86 |
+
"sae_top_5_test_accuracy": 0.859,
|
87 |
+
"sae_top_10_test_accuracy": null,
|
88 |
+
"sae_top_20_test_accuracy": null,
|
89 |
+
"sae_top_50_test_accuracy": null,
|
90 |
+
"sae_top_100_test_accuracy": null
|
91 |
+
},
|
92 |
+
{
|
93 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
94 |
+
"llm_test_accuracy": 0.9332000494003296,
|
95 |
+
"llm_top_1_test_accuracy": 0.6884,
|
96 |
+
"llm_top_2_test_accuracy": 0.737,
|
97 |
+
"llm_top_5_test_accuracy": 0.7492,
|
98 |
+
"llm_top_10_test_accuracy": null,
|
99 |
+
"llm_top_20_test_accuracy": null,
|
100 |
+
"llm_top_50_test_accuracy": null,
|
101 |
+
"llm_top_100_test_accuracy": null,
|
102 |
+
"sae_test_accuracy": 0.9212000370025635,
|
103 |
+
"sae_top_1_test_accuracy": 0.7708,
|
104 |
+
"sae_top_2_test_accuracy": 0.8034000000000001,
|
105 |
+
"sae_top_5_test_accuracy": 0.8343999999999999,
|
106 |
+
"sae_top_10_test_accuracy": null,
|
107 |
+
"sae_top_20_test_accuracy": null,
|
108 |
+
"sae_top_50_test_accuracy": null,
|
109 |
+
"sae_top_100_test_accuracy": null
|
110 |
+
},
|
111 |
+
{
|
112 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
113 |
+
"llm_test_accuracy": 0.9164000511169433,
|
114 |
+
"llm_top_1_test_accuracy": 0.598,
|
115 |
+
"llm_top_2_test_accuracy": 0.6406,
|
116 |
+
"llm_top_5_test_accuracy": 0.6748000000000001,
|
117 |
+
"llm_top_10_test_accuracy": null,
|
118 |
+
"llm_top_20_test_accuracy": null,
|
119 |
+
"llm_top_50_test_accuracy": null,
|
120 |
+
"llm_top_100_test_accuracy": null,
|
121 |
+
"sae_test_accuracy": 0.909600031375885,
|
122 |
+
"sae_top_1_test_accuracy": 0.712,
|
123 |
+
"sae_top_2_test_accuracy": 0.7582,
|
124 |
+
"sae_top_5_test_accuracy": 0.8068,
|
125 |
+
"sae_top_10_test_accuracy": null,
|
126 |
+
"sae_top_20_test_accuracy": null,
|
127 |
+
"sae_top_50_test_accuracy": null,
|
128 |
+
"sae_top_100_test_accuracy": null
|
129 |
+
},
|
130 |
+
{
|
131 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
132 |
+
"llm_test_accuracy": 0.9820000529289246,
|
133 |
+
"llm_top_1_test_accuracy": 0.672,
|
134 |
+
"llm_top_2_test_accuracy": 0.724,
|
135 |
+
"llm_top_5_test_accuracy": 0.766,
|
136 |
+
"llm_top_10_test_accuracy": null,
|
137 |
+
"llm_top_20_test_accuracy": null,
|
138 |
+
"llm_top_50_test_accuracy": null,
|
139 |
+
"llm_top_100_test_accuracy": null,
|
140 |
+
"sae_test_accuracy": 0.9665000438690186,
|
141 |
+
"sae_top_1_test_accuracy": 0.754,
|
142 |
+
"sae_top_2_test_accuracy": 0.786,
|
143 |
+
"sae_top_5_test_accuracy": 0.92,
|
144 |
+
"sae_top_10_test_accuracy": null,
|
145 |
+
"sae_top_20_test_accuracy": null,
|
146 |
+
"sae_top_50_test_accuracy": null,
|
147 |
+
"sae_top_100_test_accuracy": null
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"dataset_name": "codeparrot/github-code_results",
|
151 |
+
"llm_test_accuracy": 0.9672000408172607,
|
152 |
+
"llm_top_1_test_accuracy": 0.6668000000000001,
|
153 |
+
"llm_top_2_test_accuracy": 0.6896,
|
154 |
+
"llm_top_5_test_accuracy": 0.7656000000000001,
|
155 |
+
"llm_top_10_test_accuracy": null,
|
156 |
+
"llm_top_20_test_accuracy": null,
|
157 |
+
"llm_top_50_test_accuracy": null,
|
158 |
+
"llm_top_100_test_accuracy": null,
|
159 |
+
"sae_test_accuracy": 0.9672000527381897,
|
160 |
+
"sae_top_1_test_accuracy": 0.6784,
|
161 |
+
"sae_top_2_test_accuracy": 0.7518,
|
162 |
+
"sae_top_5_test_accuracy": 0.8684,
|
163 |
+
"sae_top_10_test_accuracy": null,
|
164 |
+
"sae_top_20_test_accuracy": null,
|
165 |
+
"sae_top_50_test_accuracy": null,
|
166 |
+
"sae_top_100_test_accuracy": null
|
167 |
+
},
|
168 |
+
{
|
169 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
170 |
+
"llm_test_accuracy": 0.9507500380277634,
|
171 |
+
"llm_top_1_test_accuracy": 0.63625,
|
172 |
+
"llm_top_2_test_accuracy": 0.7655000000000001,
|
173 |
+
"llm_top_5_test_accuracy": 0.82475,
|
174 |
+
"llm_top_10_test_accuracy": null,
|
175 |
+
"llm_top_20_test_accuracy": null,
|
176 |
+
"llm_top_50_test_accuracy": null,
|
177 |
+
"llm_top_100_test_accuracy": null,
|
178 |
+
"sae_test_accuracy": 0.9485000371932983,
|
179 |
+
"sae_top_1_test_accuracy": 0.8232499999999999,
|
180 |
+
"sae_top_2_test_accuracy": 0.8647499999999999,
|
181 |
+
"sae_top_5_test_accuracy": 0.88225,
|
182 |
+
"sae_top_10_test_accuracy": null,
|
183 |
+
"sae_top_20_test_accuracy": null,
|
184 |
+
"sae_top_50_test_accuracy": null,
|
185 |
+
"sae_top_100_test_accuracy": null
|
186 |
+
},
|
187 |
+
{
|
188 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
189 |
+
"llm_test_accuracy": 0.9994000434875489,
|
190 |
+
"llm_top_1_test_accuracy": 0.737,
|
191 |
+
"llm_top_2_test_accuracy": 0.7796,
|
192 |
+
"llm_top_5_test_accuracy": 0.9126,
|
193 |
+
"llm_top_10_test_accuracy": null,
|
194 |
+
"llm_top_20_test_accuracy": null,
|
195 |
+
"llm_top_50_test_accuracy": null,
|
196 |
+
"llm_top_100_test_accuracy": null,
|
197 |
+
"sae_test_accuracy": 0.9988000154495239,
|
198 |
+
"sae_top_1_test_accuracy": 0.8942,
|
199 |
+
"sae_top_2_test_accuracy": 0.9346,
|
200 |
+
"sae_top_5_test_accuracy": 0.9949999999999999,
|
201 |
+
"sae_top_10_test_accuracy": null,
|
202 |
+
"sae_top_20_test_accuracy": null,
|
203 |
+
"sae_top_50_test_accuracy": null,
|
204 |
+
"sae_top_100_test_accuracy": null
|
205 |
+
}
|
206 |
+
],
|
207 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
208 |
+
"sae_lens_id": "custom_sae",
|
209 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1",
|
210 |
+
"sae_lens_version": "5.3.2",
|
211 |
+
"sae_cfg_dict": {
|
212 |
+
"model_name": "gemma-2-2b",
|
213 |
+
"d_in": 2304,
|
214 |
+
"d_sae": 16384,
|
215 |
+
"hook_layer": 12,
|
216 |
+
"hook_name": "blocks.12.hook_resid_post",
|
217 |
+
"context_size": null,
|
218 |
+
"hook_head_index": null,
|
219 |
+
"architecture": "matryoshka_batch_topk",
|
220 |
+
"apply_b_dec_to_input": null,
|
221 |
+
"finetuning_scaling_factor": null,
|
222 |
+
"activation_fn_str": "",
|
223 |
+
"prepend_bos": true,
|
224 |
+
"normalize_activations": "none",
|
225 |
+
"dtype": "bfloat16",
|
226 |
+
"device": "",
|
227 |
+
"dataset_path": "",
|
228 |
+
"dataset_trust_remote_code": true,
|
229 |
+
"seqpos_slice": [
|
230 |
+
null
|
231 |
+
],
|
232 |
+
"training_tokens": -100000,
|
233 |
+
"sae_lens_training_version": null,
|
234 |
+
"neuronpedia_id": null
|
235 |
+
},
|
236 |
+
"eval_result_unstructured": {
|
237 |
+
"LabHC/bias_in_bios_class_set1_results": {
|
238 |
+
"sae_test_accuracy": {
|
239 |
+
"0": 0.9430000185966492,
|
240 |
+
"1": 0.956000030040741,
|
241 |
+
"2": 0.9450000524520874,
|
242 |
+
"6": 0.9860000610351562,
|
243 |
+
"9": 0.9700000286102295
|
244 |
+
},
|
245 |
+
"llm_test_accuracy": {
|
246 |
+
"0": 0.9510000348091125,
|
247 |
+
"1": 0.9670000672340393,
|
248 |
+
"2": 0.9520000219345093,
|
249 |
+
"6": 0.9930000305175781,
|
250 |
+
"9": 0.984000027179718
|
251 |
+
},
|
252 |
+
"llm_top_1_test_accuracy": {
|
253 |
+
"0": 0.568,
|
254 |
+
"1": 0.629,
|
255 |
+
"2": 0.679,
|
256 |
+
"6": 0.791,
|
257 |
+
"9": 0.551
|
258 |
+
},
|
259 |
+
"llm_top_2_test_accuracy": {
|
260 |
+
"0": 0.585,
|
261 |
+
"1": 0.666,
|
262 |
+
"2": 0.673,
|
263 |
+
"6": 0.801,
|
264 |
+
"9": 0.712
|
265 |
+
},
|
266 |
+
"llm_top_5_test_accuracy": {
|
267 |
+
"0": 0.72,
|
268 |
+
"1": 0.707,
|
269 |
+
"2": 0.764,
|
270 |
+
"6": 0.899,
|
271 |
+
"9": 0.864
|
272 |
+
},
|
273 |
+
"sae_top_1_test_accuracy": {
|
274 |
+
"0": 0.584,
|
275 |
+
"1": 0.623,
|
276 |
+
"2": 0.866,
|
277 |
+
"6": 0.813,
|
278 |
+
"9": 0.897
|
279 |
+
},
|
280 |
+
"sae_top_2_test_accuracy": {
|
281 |
+
"0": 0.611,
|
282 |
+
"1": 0.663,
|
283 |
+
"2": 0.864,
|
284 |
+
"6": 0.977,
|
285 |
+
"9": 0.913
|
286 |
+
},
|
287 |
+
"sae_top_5_test_accuracy": {
|
288 |
+
"0": 0.746,
|
289 |
+
"1": 0.75,
|
290 |
+
"2": 0.903,
|
291 |
+
"6": 0.975,
|
292 |
+
"9": 0.941
|
293 |
+
}
|
294 |
+
},
|
295 |
+
"LabHC/bias_in_bios_class_set2_results": {
|
296 |
+
"sae_test_accuracy": {
|
297 |
+
"11": 0.956000030040741,
|
298 |
+
"13": 0.9460000395774841,
|
299 |
+
"14": 0.9450000524520874,
|
300 |
+
"18": 0.9230000376701355,
|
301 |
+
"19": 0.9550000429153442
|
302 |
+
},
|
303 |
+
"llm_test_accuracy": {
|
304 |
+
"11": 0.9650000333786011,
|
305 |
+
"13": 0.9520000219345093,
|
306 |
+
"14": 0.9550000429153442,
|
307 |
+
"18": 0.9360000491142273,
|
308 |
+
"19": 0.9640000462532043
|
309 |
+
},
|
310 |
+
"llm_top_1_test_accuracy": {
|
311 |
+
"11": 0.558,
|
312 |
+
"13": 0.673,
|
313 |
+
"14": 0.656,
|
314 |
+
"18": 0.702,
|
315 |
+
"19": 0.793
|
316 |
+
},
|
317 |
+
"llm_top_2_test_accuracy": {
|
318 |
+
"11": 0.686,
|
319 |
+
"13": 0.713,
|
320 |
+
"14": 0.687,
|
321 |
+
"18": 0.724,
|
322 |
+
"19": 0.765
|
323 |
+
},
|
324 |
+
"llm_top_5_test_accuracy": {
|
325 |
+
"11": 0.782,
|
326 |
+
"13": 0.742,
|
327 |
+
"14": 0.716,
|
328 |
+
"18": 0.725,
|
329 |
+
"19": 0.831
|
330 |
+
},
|
331 |
+
"sae_top_1_test_accuracy": {
|
332 |
+
"11": 0.857,
|
333 |
+
"13": 0.684,
|
334 |
+
"14": 0.88,
|
335 |
+
"18": 0.704,
|
336 |
+
"19": 0.842
|
337 |
+
},
|
338 |
+
"sae_top_2_test_accuracy": {
|
339 |
+
"11": 0.852,
|
340 |
+
"13": 0.662,
|
341 |
+
"14": 0.88,
|
342 |
+
"18": 0.701,
|
343 |
+
"19": 0.838
|
344 |
+
},
|
345 |
+
"sae_top_5_test_accuracy": {
|
346 |
+
"11": 0.936,
|
347 |
+
"13": 0.762,
|
348 |
+
"14": 0.869,
|
349 |
+
"18": 0.877,
|
350 |
+
"19": 0.851
|
351 |
+
}
|
352 |
+
},
|
353 |
+
"LabHC/bias_in_bios_class_set3_results": {
|
354 |
+
"sae_test_accuracy": {
|
355 |
+
"20": 0.9510000348091125,
|
356 |
+
"21": 0.9050000309944153,
|
357 |
+
"22": 0.909000039100647,
|
358 |
+
"25": 0.9590000510215759,
|
359 |
+
"26": 0.8820000290870667
|
360 |
+
},
|
361 |
+
"llm_test_accuracy": {
|
362 |
+
"20": 0.9610000252723694,
|
363 |
+
"21": 0.9270000457763672,
|
364 |
+
"22": 0.921000063419342,
|
365 |
+
"25": 0.968000054359436,
|
366 |
+
"26": 0.8890000581741333
|
367 |
+
},
|
368 |
+
"llm_top_1_test_accuracy": {
|
369 |
+
"20": 0.713,
|
370 |
+
"21": 0.766,
|
371 |
+
"22": 0.639,
|
372 |
+
"25": 0.69,
|
373 |
+
"26": 0.634
|
374 |
+
},
|
375 |
+
"llm_top_2_test_accuracy": {
|
376 |
+
"20": 0.804,
|
377 |
+
"21": 0.777,
|
378 |
+
"22": 0.68,
|
379 |
+
"25": 0.759,
|
380 |
+
"26": 0.665
|
381 |
+
},
|
382 |
+
"llm_top_5_test_accuracy": {
|
383 |
+
"20": 0.813,
|
384 |
+
"21": 0.777,
|
385 |
+
"22": 0.686,
|
386 |
+
"25": 0.798,
|
387 |
+
"26": 0.672
|
388 |
+
},
|
389 |
+
"sae_top_1_test_accuracy": {
|
390 |
+
"20": 0.899,
|
391 |
+
"21": 0.746,
|
392 |
+
"22": 0.726,
|
393 |
+
"25": 0.877,
|
394 |
+
"26": 0.606
|
395 |
+
},
|
396 |
+
"sae_top_2_test_accuracy": {
|
397 |
+
"20": 0.9,
|
398 |
+
"21": 0.775,
|
399 |
+
"22": 0.739,
|
400 |
+
"25": 0.875,
|
401 |
+
"26": 0.728
|
402 |
+
},
|
403 |
+
"sae_top_5_test_accuracy": {
|
404 |
+
"20": 0.933,
|
405 |
+
"21": 0.825,
|
406 |
+
"22": 0.749,
|
407 |
+
"25": 0.9,
|
408 |
+
"26": 0.765
|
409 |
+
}
|
410 |
+
},
|
411 |
+
"canrager/amazon_reviews_mcauley_1and5_results": {
|
412 |
+
"sae_test_accuracy": {
|
413 |
+
"1": 0.9430000185966492,
|
414 |
+
"2": 0.9310000538825989,
|
415 |
+
"3": 0.9200000166893005,
|
416 |
+
"5": 0.9010000228881836,
|
417 |
+
"6": 0.8530000448226929
|
418 |
+
},
|
419 |
+
"llm_test_accuracy": {
|
420 |
+
"1": 0.9480000734329224,
|
421 |
+
"2": 0.9350000619888306,
|
422 |
+
"3": 0.9140000343322754,
|
423 |
+
"5": 0.9110000729560852,
|
424 |
+
"6": 0.8740000128746033
|
425 |
+
},
|
426 |
+
"llm_top_1_test_accuracy": {
|
427 |
+
"1": 0.668,
|
428 |
+
"2": 0.573,
|
429 |
+
"3": 0.591,
|
430 |
+
"5": 0.568,
|
431 |
+
"6": 0.59
|
432 |
+
},
|
433 |
+
"llm_top_2_test_accuracy": {
|
434 |
+
"1": 0.745,
|
435 |
+
"2": 0.636,
|
436 |
+
"3": 0.606,
|
437 |
+
"5": 0.595,
|
438 |
+
"6": 0.621
|
439 |
+
},
|
440 |
+
"llm_top_5_test_accuracy": {
|
441 |
+
"1": 0.778,
|
442 |
+
"2": 0.643,
|
443 |
+
"3": 0.608,
|
444 |
+
"5": 0.66,
|
445 |
+
"6": 0.685
|
446 |
+
},
|
447 |
+
"sae_top_1_test_accuracy": {
|
448 |
+
"1": 0.867,
|
449 |
+
"2": 0.617,
|
450 |
+
"3": 0.704,
|
451 |
+
"5": 0.603,
|
452 |
+
"6": 0.769
|
453 |
+
},
|
454 |
+
"sae_top_2_test_accuracy": {
|
455 |
+
"1": 0.901,
|
456 |
+
"2": 0.842,
|
457 |
+
"3": 0.696,
|
458 |
+
"5": 0.597,
|
459 |
+
"6": 0.755
|
460 |
+
},
|
461 |
+
"sae_top_5_test_accuracy": {
|
462 |
+
"1": 0.914,
|
463 |
+
"2": 0.832,
|
464 |
+
"3": 0.738,
|
465 |
+
"5": 0.77,
|
466 |
+
"6": 0.78
|
467 |
+
}
|
468 |
+
},
|
469 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment_results": {
|
470 |
+
"sae_test_accuracy": {
|
471 |
+
"1.0": 0.9660000205039978,
|
472 |
+
"5.0": 0.9670000672340393
|
473 |
+
},
|
474 |
+
"llm_test_accuracy": {
|
475 |
+
"1.0": 0.9810000658035278,
|
476 |
+
"5.0": 0.9830000400543213
|
477 |
+
},
|
478 |
+
"llm_top_1_test_accuracy": {
|
479 |
+
"1.0": 0.672,
|
480 |
+
"5.0": 0.672
|
481 |
+
},
|
482 |
+
"llm_top_2_test_accuracy": {
|
483 |
+
"1.0": 0.724,
|
484 |
+
"5.0": 0.724
|
485 |
+
},
|
486 |
+
"llm_top_5_test_accuracy": {
|
487 |
+
"1.0": 0.766,
|
488 |
+
"5.0": 0.766
|
489 |
+
},
|
490 |
+
"sae_top_1_test_accuracy": {
|
491 |
+
"1.0": 0.754,
|
492 |
+
"5.0": 0.754
|
493 |
+
},
|
494 |
+
"sae_top_2_test_accuracy": {
|
495 |
+
"1.0": 0.786,
|
496 |
+
"5.0": 0.786
|
497 |
+
},
|
498 |
+
"sae_top_5_test_accuracy": {
|
499 |
+
"1.0": 0.92,
|
500 |
+
"5.0": 0.92
|
501 |
+
}
|
502 |
+
},
|
503 |
+
"codeparrot/github-code_results": {
|
504 |
+
"sae_test_accuracy": {
|
505 |
+
"C": 0.9530000686645508,
|
506 |
+
"Python": 0.9830000400543213,
|
507 |
+
"HTML": 0.9850000739097595,
|
508 |
+
"Java": 0.9600000381469727,
|
509 |
+
"PHP": 0.9550000429153442
|
510 |
+
},
|
511 |
+
"llm_test_accuracy": {
|
512 |
+
"C": 0.9510000348091125,
|
513 |
+
"Python": 0.987000048160553,
|
514 |
+
"HTML": 0.984000027179718,
|
515 |
+
"Java": 0.968000054359436,
|
516 |
+
"PHP": 0.9460000395774841
|
517 |
+
},
|
518 |
+
"llm_top_1_test_accuracy": {
|
519 |
+
"C": 0.655,
|
520 |
+
"Python": 0.661,
|
521 |
+
"HTML": 0.802,
|
522 |
+
"Java": 0.622,
|
523 |
+
"PHP": 0.594
|
524 |
+
},
|
525 |
+
"llm_top_2_test_accuracy": {
|
526 |
+
"C": 0.667,
|
527 |
+
"Python": 0.668,
|
528 |
+
"HTML": 0.798,
|
529 |
+
"Java": 0.674,
|
530 |
+
"PHP": 0.641
|
531 |
+
},
|
532 |
+
"llm_top_5_test_accuracy": {
|
533 |
+
"C": 0.78,
|
534 |
+
"Python": 0.73,
|
535 |
+
"HTML": 0.906,
|
536 |
+
"Java": 0.728,
|
537 |
+
"PHP": 0.684
|
538 |
+
},
|
539 |
+
"sae_top_1_test_accuracy": {
|
540 |
+
"C": 0.595,
|
541 |
+
"Python": 0.635,
|
542 |
+
"HTML": 0.924,
|
543 |
+
"Java": 0.636,
|
544 |
+
"PHP": 0.602
|
545 |
+
},
|
546 |
+
"sae_top_2_test_accuracy": {
|
547 |
+
"C": 0.635,
|
548 |
+
"Python": 0.662,
|
549 |
+
"HTML": 0.929,
|
550 |
+
"Java": 0.646,
|
551 |
+
"PHP": 0.887
|
552 |
+
},
|
553 |
+
"sae_top_5_test_accuracy": {
|
554 |
+
"C": 0.821,
|
555 |
+
"Python": 0.955,
|
556 |
+
"HTML": 0.951,
|
557 |
+
"Java": 0.718,
|
558 |
+
"PHP": 0.897
|
559 |
+
}
|
560 |
+
},
|
561 |
+
"fancyzhx/ag_news_results": {
|
562 |
+
"sae_test_accuracy": {
|
563 |
+
"0": 0.9410000443458557,
|
564 |
+
"1": 0.9820000529289246,
|
565 |
+
"2": 0.9240000247955322,
|
566 |
+
"3": 0.9470000267028809
|
567 |
+
},
|
568 |
+
"llm_test_accuracy": {
|
569 |
+
"0": 0.9360000491142273,
|
570 |
+
"1": 0.9890000224113464,
|
571 |
+
"2": 0.9280000329017639,
|
572 |
+
"3": 0.9500000476837158
|
573 |
+
},
|
574 |
+
"llm_top_1_test_accuracy": {
|
575 |
+
"0": 0.58,
|
576 |
+
"1": 0.676,
|
577 |
+
"2": 0.667,
|
578 |
+
"3": 0.622
|
579 |
+
},
|
580 |
+
"llm_top_2_test_accuracy": {
|
581 |
+
"0": 0.811,
|
582 |
+
"1": 0.798,
|
583 |
+
"2": 0.693,
|
584 |
+
"3": 0.76
|
585 |
+
},
|
586 |
+
"llm_top_5_test_accuracy": {
|
587 |
+
"0": 0.815,
|
588 |
+
"1": 0.884,
|
589 |
+
"2": 0.756,
|
590 |
+
"3": 0.844
|
591 |
+
},
|
592 |
+
"sae_top_1_test_accuracy": {
|
593 |
+
"0": 0.831,
|
594 |
+
"1": 0.853,
|
595 |
+
"2": 0.794,
|
596 |
+
"3": 0.815
|
597 |
+
},
|
598 |
+
"sae_top_2_test_accuracy": {
|
599 |
+
"0": 0.827,
|
600 |
+
"1": 0.962,
|
601 |
+
"2": 0.841,
|
602 |
+
"3": 0.829
|
603 |
+
},
|
604 |
+
"sae_top_5_test_accuracy": {
|
605 |
+
"0": 0.863,
|
606 |
+
"1": 0.969,
|
607 |
+
"2": 0.844,
|
608 |
+
"3": 0.853
|
609 |
+
}
|
610 |
+
},
|
611 |
+
"Helsinki-NLP/europarl_results": {
|
612 |
+
"sae_test_accuracy": {
|
613 |
+
"en": 0.9980000257492065,
|
614 |
+
"fr": 1.0,
|
615 |
+
"de": 1.0,
|
616 |
+
"es": 0.9980000257492065,
|
617 |
+
"nl": 0.9980000257492065
|
618 |
+
},
|
619 |
+
"llm_test_accuracy": {
|
620 |
+
"en": 1.0,
|
621 |
+
"fr": 0.999000072479248,
|
622 |
+
"de": 0.999000072479248,
|
623 |
+
"es": 1.0,
|
624 |
+
"nl": 0.999000072479248
|
625 |
+
},
|
626 |
+
"llm_top_1_test_accuracy": {
|
627 |
+
"en": 0.748,
|
628 |
+
"fr": 0.615,
|
629 |
+
"de": 0.761,
|
630 |
+
"es": 0.904,
|
631 |
+
"nl": 0.657
|
632 |
+
},
|
633 |
+
"llm_top_2_test_accuracy": {
|
634 |
+
"en": 0.829,
|
635 |
+
"fr": 0.583,
|
636 |
+
"de": 0.829,
|
637 |
+
"es": 0.914,
|
638 |
+
"nl": 0.743
|
639 |
+
},
|
640 |
+
"llm_top_5_test_accuracy": {
|
641 |
+
"en": 0.892,
|
642 |
+
"fr": 0.914,
|
643 |
+
"de": 0.917,
|
644 |
+
"es": 0.984,
|
645 |
+
"nl": 0.856
|
646 |
+
},
|
647 |
+
"sae_top_1_test_accuracy": {
|
648 |
+
"en": 0.997,
|
649 |
+
"fr": 0.989,
|
650 |
+
"de": 0.731,
|
651 |
+
"es": 0.994,
|
652 |
+
"nl": 0.76
|
653 |
+
},
|
654 |
+
"sae_top_2_test_accuracy": {
|
655 |
+
"en": 1.0,
|
656 |
+
"fr": 0.992,
|
657 |
+
"de": 0.91,
|
658 |
+
"es": 0.994,
|
659 |
+
"nl": 0.777
|
660 |
+
},
|
661 |
+
"sae_top_5_test_accuracy": {
|
662 |
+
"en": 0.999,
|
663 |
+
"fr": 0.995,
|
664 |
+
"de": 0.987,
|
665 |
+
"es": 0.995,
|
666 |
+
"nl": 0.999
|
667 |
+
}
|
668 |
+
}
|
669 |
+
}
|
670 |
+
}
|
sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,670 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "sparse_probing",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"LabHC/bias_in_bios_class_set2",
|
8 |
+
"LabHC/bias_in_bios_class_set3",
|
9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
11 |
+
"codeparrot/github-code",
|
12 |
+
"fancyzhx/ag_news",
|
13 |
+
"Helsinki-NLP/europarl"
|
14 |
+
],
|
15 |
+
"probe_train_set_size": 4000,
|
16 |
+
"probe_test_set_size": 1000,
|
17 |
+
"context_length": 128,
|
18 |
+
"sae_batch_size": 125,
|
19 |
+
"llm_batch_size": 32,
|
20 |
+
"llm_dtype": "bfloat16",
|
21 |
+
"model_name": "gemma-2-2b",
|
22 |
+
"k_values": [
|
23 |
+
1,
|
24 |
+
2,
|
25 |
+
5
|
26 |
+
],
|
27 |
+
"lower_vram_usage": false
|
28 |
+
},
|
29 |
+
"eval_id": "b3d9ede2-3e45-4bd9-bb19-61e5fe3f8eb9",
|
30 |
+
"datetime_epoch_millis": 1737823308260,
|
31 |
+
"eval_result_metrics": {
|
32 |
+
"llm": {
|
33 |
+
"llm_test_accuracy": 0.9590937938541175,
|
34 |
+
"llm_top_1_test_accuracy": 0.66480625,
|
35 |
+
"llm_top_2_test_accuracy": 0.7173375000000001,
|
36 |
+
"llm_top_5_test_accuracy": 0.78036875,
|
37 |
+
"llm_top_10_test_accuracy": null,
|
38 |
+
"llm_top_20_test_accuracy": null,
|
39 |
+
"llm_top_50_test_accuracy": null,
|
40 |
+
"llm_top_100_test_accuracy": null
|
41 |
+
},
|
42 |
+
"sae": {
|
43 |
+
"sae_test_accuracy": 0.9559312954545021,
|
44 |
+
"sae_top_1_test_accuracy": 0.7493249999999999,
|
45 |
+
"sae_top_2_test_accuracy": 0.8200375,
|
46 |
+
"sae_top_5_test_accuracy": 0.87965,
|
47 |
+
"sae_top_10_test_accuracy": null,
|
48 |
+
"sae_top_20_test_accuracy": null,
|
49 |
+
"sae_top_50_test_accuracy": null,
|
50 |
+
"sae_top_100_test_accuracy": null
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"eval_result_details": [
|
54 |
+
{
|
55 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
56 |
+
"llm_test_accuracy": 0.9694000363349915,
|
57 |
+
"llm_top_1_test_accuracy": 0.6436000000000001,
|
58 |
+
"llm_top_2_test_accuracy": 0.6874,
|
59 |
+
"llm_top_5_test_accuracy": 0.7908,
|
60 |
+
"llm_top_10_test_accuracy": null,
|
61 |
+
"llm_top_20_test_accuracy": null,
|
62 |
+
"llm_top_50_test_accuracy": null,
|
63 |
+
"llm_top_100_test_accuracy": null,
|
64 |
+
"sae_test_accuracy": 0.962000036239624,
|
65 |
+
"sae_top_1_test_accuracy": 0.7726,
|
66 |
+
"sae_top_2_test_accuracy": 0.8816,
|
67 |
+
"sae_top_5_test_accuracy": 0.8981999999999999,
|
68 |
+
"sae_top_10_test_accuracy": null,
|
69 |
+
"sae_top_20_test_accuracy": null,
|
70 |
+
"sae_top_50_test_accuracy": null,
|
71 |
+
"sae_top_100_test_accuracy": null
|
72 |
+
},
|
73 |
+
{
|
74 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
75 |
+
"llm_test_accuracy": 0.9544000387191772,
|
76 |
+
"llm_top_1_test_accuracy": 0.6764,
|
77 |
+
"llm_top_2_test_accuracy": 0.7150000000000001,
|
78 |
+
"llm_top_5_test_accuracy": 0.7592000000000001,
|
79 |
+
"llm_top_10_test_accuracy": null,
|
80 |
+
"llm_top_20_test_accuracy": null,
|
81 |
+
"llm_top_50_test_accuracy": null,
|
82 |
+
"llm_top_100_test_accuracy": null,
|
83 |
+
"sae_test_accuracy": 0.9470000505447388,
|
84 |
+
"sae_top_1_test_accuracy": 0.6846,
|
85 |
+
"sae_top_2_test_accuracy": 0.8221999999999999,
|
86 |
+
"sae_top_5_test_accuracy": 0.8493999999999999,
|
87 |
+
"sae_top_10_test_accuracy": null,
|
88 |
+
"sae_top_20_test_accuracy": null,
|
89 |
+
"sae_top_50_test_accuracy": null,
|
90 |
+
"sae_top_100_test_accuracy": null
|
91 |
+
},
|
92 |
+
{
|
93 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
94 |
+
"llm_test_accuracy": 0.9332000494003296,
|
95 |
+
"llm_top_1_test_accuracy": 0.6884,
|
96 |
+
"llm_top_2_test_accuracy": 0.737,
|
97 |
+
"llm_top_5_test_accuracy": 0.7492,
|
98 |
+
"llm_top_10_test_accuracy": null,
|
99 |
+
"llm_top_20_test_accuracy": null,
|
100 |
+
"llm_top_50_test_accuracy": null,
|
101 |
+
"llm_top_100_test_accuracy": null,
|
102 |
+
"sae_test_accuracy": 0.9270000338554383,
|
103 |
+
"sae_top_1_test_accuracy": 0.6412000000000001,
|
104 |
+
"sae_top_2_test_accuracy": 0.7872,
|
105 |
+
"sae_top_5_test_accuracy": 0.8530000000000001,
|
106 |
+
"sae_top_10_test_accuracy": null,
|
107 |
+
"sae_top_20_test_accuracy": null,
|
108 |
+
"sae_top_50_test_accuracy": null,
|
109 |
+
"sae_top_100_test_accuracy": null
|
110 |
+
},
|
111 |
+
{
|
112 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
113 |
+
"llm_test_accuracy": 0.9164000511169433,
|
114 |
+
"llm_top_1_test_accuracy": 0.598,
|
115 |
+
"llm_top_2_test_accuracy": 0.6406,
|
116 |
+
"llm_top_5_test_accuracy": 0.6748000000000001,
|
117 |
+
"llm_top_10_test_accuracy": null,
|
118 |
+
"llm_top_20_test_accuracy": null,
|
119 |
+
"llm_top_50_test_accuracy": null,
|
120 |
+
"llm_top_100_test_accuracy": null,
|
121 |
+
"sae_test_accuracy": 0.9194000363349915,
|
122 |
+
"sae_top_1_test_accuracy": 0.7292,
|
123 |
+
"sae_top_2_test_accuracy": 0.7478,
|
124 |
+
"sae_top_5_test_accuracy": 0.8224,
|
125 |
+
"sae_top_10_test_accuracy": null,
|
126 |
+
"sae_top_20_test_accuracy": null,
|
127 |
+
"sae_top_50_test_accuracy": null,
|
128 |
+
"sae_top_100_test_accuracy": null
|
129 |
+
},
|
130 |
+
{
|
131 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
132 |
+
"llm_test_accuracy": 0.9820000529289246,
|
133 |
+
"llm_top_1_test_accuracy": 0.672,
|
134 |
+
"llm_top_2_test_accuracy": 0.724,
|
135 |
+
"llm_top_5_test_accuracy": 0.766,
|
136 |
+
"llm_top_10_test_accuracy": null,
|
137 |
+
"llm_top_20_test_accuracy": null,
|
138 |
+
"llm_top_50_test_accuracy": null,
|
139 |
+
"llm_top_100_test_accuracy": null,
|
140 |
+
"sae_test_accuracy": 0.9720000624656677,
|
141 |
+
"sae_top_1_test_accuracy": 0.852,
|
142 |
+
"sae_top_2_test_accuracy": 0.877,
|
143 |
+
"sae_top_5_test_accuracy": 0.904,
|
144 |
+
"sae_top_10_test_accuracy": null,
|
145 |
+
"sae_top_20_test_accuracy": null,
|
146 |
+
"sae_top_50_test_accuracy": null,
|
147 |
+
"sae_top_100_test_accuracy": null
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"dataset_name": "codeparrot/github-code_results",
|
151 |
+
"llm_test_accuracy": 0.9672000408172607,
|
152 |
+
"llm_top_1_test_accuracy": 0.6668000000000001,
|
153 |
+
"llm_top_2_test_accuracy": 0.6896,
|
154 |
+
"llm_top_5_test_accuracy": 0.7656000000000001,
|
155 |
+
"llm_top_10_test_accuracy": null,
|
156 |
+
"llm_top_20_test_accuracy": null,
|
157 |
+
"llm_top_50_test_accuracy": null,
|
158 |
+
"llm_top_100_test_accuracy": null,
|
159 |
+
"sae_test_accuracy": 0.9676000356674195,
|
160 |
+
"sae_top_1_test_accuracy": 0.6361999999999999,
|
161 |
+
"sae_top_2_test_accuracy": 0.7402,
|
162 |
+
"sae_top_5_test_accuracy": 0.8305999999999999,
|
163 |
+
"sae_top_10_test_accuracy": null,
|
164 |
+
"sae_top_20_test_accuracy": null,
|
165 |
+
"sae_top_50_test_accuracy": null,
|
166 |
+
"sae_top_100_test_accuracy": null
|
167 |
+
},
|
168 |
+
{
|
169 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
170 |
+
"llm_test_accuracy": 0.9507500380277634,
|
171 |
+
"llm_top_1_test_accuracy": 0.63625,
|
172 |
+
"llm_top_2_test_accuracy": 0.7655000000000001,
|
173 |
+
"llm_top_5_test_accuracy": 0.82475,
|
174 |
+
"llm_top_10_test_accuracy": null,
|
175 |
+
"llm_top_20_test_accuracy": null,
|
176 |
+
"llm_top_50_test_accuracy": null,
|
177 |
+
"llm_top_100_test_accuracy": null,
|
178 |
+
"sae_test_accuracy": 0.9532500505447388,
|
179 |
+
"sae_top_1_test_accuracy": 0.815,
|
180 |
+
"sae_top_2_test_accuracy": 0.8205,
|
181 |
+
"sae_top_5_test_accuracy": 0.892,
|
182 |
+
"sae_top_10_test_accuracy": null,
|
183 |
+
"sae_top_20_test_accuracy": null,
|
184 |
+
"sae_top_50_test_accuracy": null,
|
185 |
+
"sae_top_100_test_accuracy": null
|
186 |
+
},
|
187 |
+
{
|
188 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
189 |
+
"llm_test_accuracy": 0.9994000434875489,
|
190 |
+
"llm_top_1_test_accuracy": 0.737,
|
191 |
+
"llm_top_2_test_accuracy": 0.7796,
|
192 |
+
"llm_top_5_test_accuracy": 0.9126,
|
193 |
+
"llm_top_10_test_accuracy": null,
|
194 |
+
"llm_top_20_test_accuracy": null,
|
195 |
+
"llm_top_50_test_accuracy": null,
|
196 |
+
"llm_top_100_test_accuracy": null,
|
197 |
+
"sae_test_accuracy": 0.9992000579833984,
|
198 |
+
"sae_top_1_test_accuracy": 0.8638,
|
199 |
+
"sae_top_2_test_accuracy": 0.8837999999999999,
|
200 |
+
"sae_top_5_test_accuracy": 0.9876000000000001,
|
201 |
+
"sae_top_10_test_accuracy": null,
|
202 |
+
"sae_top_20_test_accuracy": null,
|
203 |
+
"sae_top_50_test_accuracy": null,
|
204 |
+
"sae_top_100_test_accuracy": null
|
205 |
+
}
|
206 |
+
],
|
207 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
208 |
+
"sae_lens_id": "custom_sae",
|
209 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2",
|
210 |
+
"sae_lens_version": "5.3.2",
|
211 |
+
"sae_cfg_dict": {
|
212 |
+
"model_name": "gemma-2-2b",
|
213 |
+
"d_in": 2304,
|
214 |
+
"d_sae": 16384,
|
215 |
+
"hook_layer": 12,
|
216 |
+
"hook_name": "blocks.12.hook_resid_post",
|
217 |
+
"context_size": null,
|
218 |
+
"hook_head_index": null,
|
219 |
+
"architecture": "matryoshka_batch_topk",
|
220 |
+
"apply_b_dec_to_input": null,
|
221 |
+
"finetuning_scaling_factor": null,
|
222 |
+
"activation_fn_str": "",
|
223 |
+
"prepend_bos": true,
|
224 |
+
"normalize_activations": "none",
|
225 |
+
"dtype": "bfloat16",
|
226 |
+
"device": "",
|
227 |
+
"dataset_path": "",
|
228 |
+
"dataset_trust_remote_code": true,
|
229 |
+
"seqpos_slice": [
|
230 |
+
null
|
231 |
+
],
|
232 |
+
"training_tokens": -100000,
|
233 |
+
"sae_lens_training_version": null,
|
234 |
+
"neuronpedia_id": null
|
235 |
+
},
|
236 |
+
"eval_result_unstructured": {
|
237 |
+
"LabHC/bias_in_bios_class_set1_results": {
|
238 |
+
"sae_test_accuracy": {
|
239 |
+
"0": 0.9450000524520874,
|
240 |
+
"1": 0.9570000171661377,
|
241 |
+
"2": 0.9510000348091125,
|
242 |
+
"6": 0.9880000352859497,
|
243 |
+
"9": 0.9690000414848328
|
244 |
+
},
|
245 |
+
"llm_test_accuracy": {
|
246 |
+
"0": 0.9510000348091125,
|
247 |
+
"1": 0.9670000672340393,
|
248 |
+
"2": 0.9520000219345093,
|
249 |
+
"6": 0.9930000305175781,
|
250 |
+
"9": 0.984000027179718
|
251 |
+
},
|
252 |
+
"llm_top_1_test_accuracy": {
|
253 |
+
"0": 0.568,
|
254 |
+
"1": 0.629,
|
255 |
+
"2": 0.679,
|
256 |
+
"6": 0.791,
|
257 |
+
"9": 0.551
|
258 |
+
},
|
259 |
+
"llm_top_2_test_accuracy": {
|
260 |
+
"0": 0.585,
|
261 |
+
"1": 0.666,
|
262 |
+
"2": 0.673,
|
263 |
+
"6": 0.801,
|
264 |
+
"9": 0.712
|
265 |
+
},
|
266 |
+
"llm_top_5_test_accuracy": {
|
267 |
+
"0": 0.72,
|
268 |
+
"1": 0.707,
|
269 |
+
"2": 0.764,
|
270 |
+
"6": 0.899,
|
271 |
+
"9": 0.864
|
272 |
+
},
|
273 |
+
"sae_top_1_test_accuracy": {
|
274 |
+
"0": 0.566,
|
275 |
+
"1": 0.635,
|
276 |
+
"2": 0.896,
|
277 |
+
"6": 0.822,
|
278 |
+
"9": 0.944
|
279 |
+
},
|
280 |
+
"sae_top_2_test_accuracy": {
|
281 |
+
"0": 0.821,
|
282 |
+
"1": 0.777,
|
283 |
+
"2": 0.898,
|
284 |
+
"6": 0.97,
|
285 |
+
"9": 0.942
|
286 |
+
},
|
287 |
+
"sae_top_5_test_accuracy": {
|
288 |
+
"0": 0.824,
|
289 |
+
"1": 0.841,
|
290 |
+
"2": 0.904,
|
291 |
+
"6": 0.976,
|
292 |
+
"9": 0.946
|
293 |
+
}
|
294 |
+
},
|
295 |
+
"LabHC/bias_in_bios_class_set2_results": {
|
296 |
+
"sae_test_accuracy": {
|
297 |
+
"11": 0.9620000720024109,
|
298 |
+
"13": 0.9460000395774841,
|
299 |
+
"14": 0.9470000267028809,
|
300 |
+
"18": 0.9250000715255737,
|
301 |
+
"19": 0.9550000429153442
|
302 |
+
},
|
303 |
+
"llm_test_accuracy": {
|
304 |
+
"11": 0.9650000333786011,
|
305 |
+
"13": 0.9520000219345093,
|
306 |
+
"14": 0.9550000429153442,
|
307 |
+
"18": 0.9360000491142273,
|
308 |
+
"19": 0.9640000462532043
|
309 |
+
},
|
310 |
+
"llm_top_1_test_accuracy": {
|
311 |
+
"11": 0.558,
|
312 |
+
"13": 0.673,
|
313 |
+
"14": 0.656,
|
314 |
+
"18": 0.702,
|
315 |
+
"19": 0.793
|
316 |
+
},
|
317 |
+
"llm_top_2_test_accuracy": {
|
318 |
+
"11": 0.686,
|
319 |
+
"13": 0.713,
|
320 |
+
"14": 0.687,
|
321 |
+
"18": 0.724,
|
322 |
+
"19": 0.765
|
323 |
+
},
|
324 |
+
"llm_top_5_test_accuracy": {
|
325 |
+
"11": 0.782,
|
326 |
+
"13": 0.742,
|
327 |
+
"14": 0.716,
|
328 |
+
"18": 0.725,
|
329 |
+
"19": 0.831
|
330 |
+
},
|
331 |
+
"sae_top_1_test_accuracy": {
|
332 |
+
"11": 0.567,
|
333 |
+
"13": 0.672,
|
334 |
+
"14": 0.663,
|
335 |
+
"18": 0.713,
|
336 |
+
"19": 0.808
|
337 |
+
},
|
338 |
+
"sae_top_2_test_accuracy": {
|
339 |
+
"11": 0.854,
|
340 |
+
"13": 0.786,
|
341 |
+
"14": 0.88,
|
342 |
+
"18": 0.745,
|
343 |
+
"19": 0.846
|
344 |
+
},
|
345 |
+
"sae_top_5_test_accuracy": {
|
346 |
+
"11": 0.875,
|
347 |
+
"13": 0.787,
|
348 |
+
"14": 0.886,
|
349 |
+
"18": 0.814,
|
350 |
+
"19": 0.885
|
351 |
+
}
|
352 |
+
},
|
353 |
+
"LabHC/bias_in_bios_class_set3_results": {
|
354 |
+
"sae_test_accuracy": {
|
355 |
+
"20": 0.9590000510215759,
|
356 |
+
"21": 0.9190000295639038,
|
357 |
+
"22": 0.906000018119812,
|
358 |
+
"25": 0.9640000462532043,
|
359 |
+
"26": 0.8870000243186951
|
360 |
+
},
|
361 |
+
"llm_test_accuracy": {
|
362 |
+
"20": 0.9610000252723694,
|
363 |
+
"21": 0.9270000457763672,
|
364 |
+
"22": 0.921000063419342,
|
365 |
+
"25": 0.968000054359436,
|
366 |
+
"26": 0.8890000581741333
|
367 |
+
},
|
368 |
+
"llm_top_1_test_accuracy": {
|
369 |
+
"20": 0.713,
|
370 |
+
"21": 0.766,
|
371 |
+
"22": 0.639,
|
372 |
+
"25": 0.69,
|
373 |
+
"26": 0.634
|
374 |
+
},
|
375 |
+
"llm_top_2_test_accuracy": {
|
376 |
+
"20": 0.804,
|
377 |
+
"21": 0.777,
|
378 |
+
"22": 0.68,
|
379 |
+
"25": 0.759,
|
380 |
+
"26": 0.665
|
381 |
+
},
|
382 |
+
"llm_top_5_test_accuracy": {
|
383 |
+
"20": 0.813,
|
384 |
+
"21": 0.777,
|
385 |
+
"22": 0.686,
|
386 |
+
"25": 0.798,
|
387 |
+
"26": 0.672
|
388 |
+
},
|
389 |
+
"sae_top_1_test_accuracy": {
|
390 |
+
"20": 0.883,
|
391 |
+
"21": 0.503,
|
392 |
+
"22": 0.467,
|
393 |
+
"25": 0.726,
|
394 |
+
"26": 0.627
|
395 |
+
},
|
396 |
+
"sae_top_2_test_accuracy": {
|
397 |
+
"20": 0.881,
|
398 |
+
"21": 0.738,
|
399 |
+
"22": 0.74,
|
400 |
+
"25": 0.861,
|
401 |
+
"26": 0.716
|
402 |
+
},
|
403 |
+
"sae_top_5_test_accuracy": {
|
404 |
+
"20": 0.913,
|
405 |
+
"21": 0.852,
|
406 |
+
"22": 0.863,
|
407 |
+
"25": 0.87,
|
408 |
+
"26": 0.767
|
409 |
+
}
|
410 |
+
},
|
411 |
+
"canrager/amazon_reviews_mcauley_1and5_results": {
|
412 |
+
"sae_test_accuracy": {
|
413 |
+
"1": 0.9420000314712524,
|
414 |
+
"2": 0.9420000314712524,
|
415 |
+
"3": 0.9130000472068787,
|
416 |
+
"5": 0.9180000424385071,
|
417 |
+
"6": 0.8820000290870667
|
418 |
+
},
|
419 |
+
"llm_test_accuracy": {
|
420 |
+
"1": 0.9480000734329224,
|
421 |
+
"2": 0.9350000619888306,
|
422 |
+
"3": 0.9140000343322754,
|
423 |
+
"5": 0.9110000729560852,
|
424 |
+
"6": 0.8740000128746033
|
425 |
+
},
|
426 |
+
"llm_top_1_test_accuracy": {
|
427 |
+
"1": 0.668,
|
428 |
+
"2": 0.573,
|
429 |
+
"3": 0.591,
|
430 |
+
"5": 0.568,
|
431 |
+
"6": 0.59
|
432 |
+
},
|
433 |
+
"llm_top_2_test_accuracy": {
|
434 |
+
"1": 0.745,
|
435 |
+
"2": 0.636,
|
436 |
+
"3": 0.606,
|
437 |
+
"5": 0.595,
|
438 |
+
"6": 0.621
|
439 |
+
},
|
440 |
+
"llm_top_5_test_accuracy": {
|
441 |
+
"1": 0.778,
|
442 |
+
"2": 0.643,
|
443 |
+
"3": 0.608,
|
444 |
+
"5": 0.66,
|
445 |
+
"6": 0.685
|
446 |
+
},
|
447 |
+
"sae_top_1_test_accuracy": {
|
448 |
+
"1": 0.854,
|
449 |
+
"2": 0.641,
|
450 |
+
"3": 0.641,
|
451 |
+
"5": 0.8,
|
452 |
+
"6": 0.71
|
453 |
+
},
|
454 |
+
"sae_top_2_test_accuracy": {
|
455 |
+
"1": 0.843,
|
456 |
+
"2": 0.644,
|
457 |
+
"3": 0.69,
|
458 |
+
"5": 0.84,
|
459 |
+
"6": 0.722
|
460 |
+
},
|
461 |
+
"sae_top_5_test_accuracy": {
|
462 |
+
"1": 0.921,
|
463 |
+
"2": 0.869,
|
464 |
+
"3": 0.713,
|
465 |
+
"5": 0.851,
|
466 |
+
"6": 0.758
|
467 |
+
}
|
468 |
+
},
|
469 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment_results": {
|
470 |
+
"sae_test_accuracy": {
|
471 |
+
"1.0": 0.9720000624656677,
|
472 |
+
"5.0": 0.9720000624656677
|
473 |
+
},
|
474 |
+
"llm_test_accuracy": {
|
475 |
+
"1.0": 0.9810000658035278,
|
476 |
+
"5.0": 0.9830000400543213
|
477 |
+
},
|
478 |
+
"llm_top_1_test_accuracy": {
|
479 |
+
"1.0": 0.672,
|
480 |
+
"5.0": 0.672
|
481 |
+
},
|
482 |
+
"llm_top_2_test_accuracy": {
|
483 |
+
"1.0": 0.724,
|
484 |
+
"5.0": 0.724
|
485 |
+
},
|
486 |
+
"llm_top_5_test_accuracy": {
|
487 |
+
"1.0": 0.766,
|
488 |
+
"5.0": 0.766
|
489 |
+
},
|
490 |
+
"sae_top_1_test_accuracy": {
|
491 |
+
"1.0": 0.852,
|
492 |
+
"5.0": 0.852
|
493 |
+
},
|
494 |
+
"sae_top_2_test_accuracy": {
|
495 |
+
"1.0": 0.877,
|
496 |
+
"5.0": 0.877
|
497 |
+
},
|
498 |
+
"sae_top_5_test_accuracy": {
|
499 |
+
"1.0": 0.904,
|
500 |
+
"5.0": 0.904
|
501 |
+
}
|
502 |
+
},
|
503 |
+
"codeparrot/github-code_results": {
|
504 |
+
"sae_test_accuracy": {
|
505 |
+
"C": 0.9570000171661377,
|
506 |
+
"Python": 0.9770000576972961,
|
507 |
+
"HTML": 0.987000048160553,
|
508 |
+
"Java": 0.9610000252723694,
|
509 |
+
"PHP": 0.956000030040741
|
510 |
+
},
|
511 |
+
"llm_test_accuracy": {
|
512 |
+
"C": 0.9510000348091125,
|
513 |
+
"Python": 0.987000048160553,
|
514 |
+
"HTML": 0.984000027179718,
|
515 |
+
"Java": 0.968000054359436,
|
516 |
+
"PHP": 0.9460000395774841
|
517 |
+
},
|
518 |
+
"llm_top_1_test_accuracy": {
|
519 |
+
"C": 0.655,
|
520 |
+
"Python": 0.661,
|
521 |
+
"HTML": 0.802,
|
522 |
+
"Java": 0.622,
|
523 |
+
"PHP": 0.594
|
524 |
+
},
|
525 |
+
"llm_top_2_test_accuracy": {
|
526 |
+
"C": 0.667,
|
527 |
+
"Python": 0.668,
|
528 |
+
"HTML": 0.798,
|
529 |
+
"Java": 0.674,
|
530 |
+
"PHP": 0.641
|
531 |
+
},
|
532 |
+
"llm_top_5_test_accuracy": {
|
533 |
+
"C": 0.78,
|
534 |
+
"Python": 0.73,
|
535 |
+
"HTML": 0.906,
|
536 |
+
"Java": 0.728,
|
537 |
+
"PHP": 0.684
|
538 |
+
},
|
539 |
+
"sae_top_1_test_accuracy": {
|
540 |
+
"C": 0.614,
|
541 |
+
"Python": 0.629,
|
542 |
+
"HTML": 0.706,
|
543 |
+
"Java": 0.643,
|
544 |
+
"PHP": 0.589
|
545 |
+
},
|
546 |
+
"sae_top_2_test_accuracy": {
|
547 |
+
"C": 0.638,
|
548 |
+
"Python": 0.665,
|
549 |
+
"HTML": 0.936,
|
550 |
+
"Java": 0.652,
|
551 |
+
"PHP": 0.81
|
552 |
+
},
|
553 |
+
"sae_top_5_test_accuracy": {
|
554 |
+
"C": 0.836,
|
555 |
+
"Python": 0.923,
|
556 |
+
"HTML": 0.927,
|
557 |
+
"Java": 0.658,
|
558 |
+
"PHP": 0.809
|
559 |
+
}
|
560 |
+
},
|
561 |
+
"fancyzhx/ag_news_results": {
|
562 |
+
"sae_test_accuracy": {
|
563 |
+
"0": 0.9420000314712524,
|
564 |
+
"1": 0.9850000739097595,
|
565 |
+
"2": 0.937000036239624,
|
566 |
+
"3": 0.9490000605583191
|
567 |
+
},
|
568 |
+
"llm_test_accuracy": {
|
569 |
+
"0": 0.9360000491142273,
|
570 |
+
"1": 0.9890000224113464,
|
571 |
+
"2": 0.9280000329017639,
|
572 |
+
"3": 0.9500000476837158
|
573 |
+
},
|
574 |
+
"llm_top_1_test_accuracy": {
|
575 |
+
"0": 0.58,
|
576 |
+
"1": 0.676,
|
577 |
+
"2": 0.667,
|
578 |
+
"3": 0.622
|
579 |
+
},
|
580 |
+
"llm_top_2_test_accuracy": {
|
581 |
+
"0": 0.811,
|
582 |
+
"1": 0.798,
|
583 |
+
"2": 0.693,
|
584 |
+
"3": 0.76
|
585 |
+
},
|
586 |
+
"llm_top_5_test_accuracy": {
|
587 |
+
"0": 0.815,
|
588 |
+
"1": 0.884,
|
589 |
+
"2": 0.756,
|
590 |
+
"3": 0.844
|
591 |
+
},
|
592 |
+
"sae_top_1_test_accuracy": {
|
593 |
+
"0": 0.737,
|
594 |
+
"1": 0.964,
|
595 |
+
"2": 0.868,
|
596 |
+
"3": 0.691
|
597 |
+
},
|
598 |
+
"sae_top_2_test_accuracy": {
|
599 |
+
"0": 0.76,
|
600 |
+
"1": 0.968,
|
601 |
+
"2": 0.864,
|
602 |
+
"3": 0.69
|
603 |
+
},
|
604 |
+
"sae_top_5_test_accuracy": {
|
605 |
+
"0": 0.884,
|
606 |
+
"1": 0.967,
|
607 |
+
"2": 0.867,
|
608 |
+
"3": 0.85
|
609 |
+
}
|
610 |
+
},
|
611 |
+
"Helsinki-NLP/europarl_results": {
|
612 |
+
"sae_test_accuracy": {
|
613 |
+
"en": 0.999000072479248,
|
614 |
+
"fr": 1.0,
|
615 |
+
"de": 0.999000072479248,
|
616 |
+
"es": 0.999000072479248,
|
617 |
+
"nl": 0.999000072479248
|
618 |
+
},
|
619 |
+
"llm_test_accuracy": {
|
620 |
+
"en": 1.0,
|
621 |
+
"fr": 0.999000072479248,
|
622 |
+
"de": 0.999000072479248,
|
623 |
+
"es": 1.0,
|
624 |
+
"nl": 0.999000072479248
|
625 |
+
},
|
626 |
+
"llm_top_1_test_accuracy": {
|
627 |
+
"en": 0.748,
|
628 |
+
"fr": 0.615,
|
629 |
+
"de": 0.761,
|
630 |
+
"es": 0.904,
|
631 |
+
"nl": 0.657
|
632 |
+
},
|
633 |
+
"llm_top_2_test_accuracy": {
|
634 |
+
"en": 0.829,
|
635 |
+
"fr": 0.583,
|
636 |
+
"de": 0.829,
|
637 |
+
"es": 0.914,
|
638 |
+
"nl": 0.743
|
639 |
+
},
|
640 |
+
"llm_top_5_test_accuracy": {
|
641 |
+
"en": 0.892,
|
642 |
+
"fr": 0.914,
|
643 |
+
"de": 0.917,
|
644 |
+
"es": 0.984,
|
645 |
+
"nl": 0.856
|
646 |
+
},
|
647 |
+
"sae_top_1_test_accuracy": {
|
648 |
+
"en": 0.998,
|
649 |
+
"fr": 0.804,
|
650 |
+
"de": 0.925,
|
651 |
+
"es": 0.86,
|
652 |
+
"nl": 0.732
|
653 |
+
},
|
654 |
+
"sae_top_2_test_accuracy": {
|
655 |
+
"en": 1.0,
|
656 |
+
"fr": 0.84,
|
657 |
+
"de": 0.925,
|
658 |
+
"es": 0.877,
|
659 |
+
"nl": 0.777
|
660 |
+
},
|
661 |
+
"sae_top_5_test_accuracy": {
|
662 |
+
"en": 0.999,
|
663 |
+
"fr": 0.989,
|
664 |
+
"de": 0.962,
|
665 |
+
"es": 0.992,
|
666 |
+
"nl": 0.996
|
667 |
+
}
|
668 |
+
}
|
669 |
+
}
|
670 |
+
}
|
sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,670 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "sparse_probing",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"LabHC/bias_in_bios_class_set2",
|
8 |
+
"LabHC/bias_in_bios_class_set3",
|
9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
11 |
+
"codeparrot/github-code",
|
12 |
+
"fancyzhx/ag_news",
|
13 |
+
"Helsinki-NLP/europarl"
|
14 |
+
],
|
15 |
+
"probe_train_set_size": 4000,
|
16 |
+
"probe_test_set_size": 1000,
|
17 |
+
"context_length": 128,
|
18 |
+
"sae_batch_size": 125,
|
19 |
+
"llm_batch_size": 32,
|
20 |
+
"llm_dtype": "bfloat16",
|
21 |
+
"model_name": "gemma-2-2b",
|
22 |
+
"k_values": [
|
23 |
+
1,
|
24 |
+
2,
|
25 |
+
5
|
26 |
+
],
|
27 |
+
"lower_vram_usage": false
|
28 |
+
},
|
29 |
+
"eval_id": "729f557d-8734-49c0-aae2-4755425362b9",
|
30 |
+
"datetime_epoch_millis": 1737823633669,
|
31 |
+
"eval_result_metrics": {
|
32 |
+
"llm": {
|
33 |
+
"llm_test_accuracy": 0.9590937938541175,
|
34 |
+
"llm_top_1_test_accuracy": 0.66480625,
|
35 |
+
"llm_top_2_test_accuracy": 0.7173375000000001,
|
36 |
+
"llm_top_5_test_accuracy": 0.78036875,
|
37 |
+
"llm_top_10_test_accuracy": null,
|
38 |
+
"llm_top_20_test_accuracy": null,
|
39 |
+
"llm_top_50_test_accuracy": null,
|
40 |
+
"llm_top_100_test_accuracy": null
|
41 |
+
},
|
42 |
+
"sae": {
|
43 |
+
"sae_test_accuracy": 0.9556812960654497,
|
44 |
+
"sae_top_1_test_accuracy": 0.7781562499999999,
|
45 |
+
"sae_top_2_test_accuracy": 0.820825,
|
46 |
+
"sae_top_5_test_accuracy": 0.8754812500000001,
|
47 |
+
"sae_top_10_test_accuracy": null,
|
48 |
+
"sae_top_20_test_accuracy": null,
|
49 |
+
"sae_top_50_test_accuracy": null,
|
50 |
+
"sae_top_100_test_accuracy": null
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"eval_result_details": [
|
54 |
+
{
|
55 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
56 |
+
"llm_test_accuracy": 0.9694000363349915,
|
57 |
+
"llm_top_1_test_accuracy": 0.6436000000000001,
|
58 |
+
"llm_top_2_test_accuracy": 0.6874,
|
59 |
+
"llm_top_5_test_accuracy": 0.7908,
|
60 |
+
"llm_top_10_test_accuracy": null,
|
61 |
+
"llm_top_20_test_accuracy": null,
|
62 |
+
"llm_top_50_test_accuracy": null,
|
63 |
+
"llm_top_100_test_accuracy": null,
|
64 |
+
"sae_test_accuracy": 0.9664000391960144,
|
65 |
+
"sae_top_1_test_accuracy": 0.7215999999999999,
|
66 |
+
"sae_top_2_test_accuracy": 0.7818,
|
67 |
+
"sae_top_5_test_accuracy": 0.876,
|
68 |
+
"sae_top_10_test_accuracy": null,
|
69 |
+
"sae_top_20_test_accuracy": null,
|
70 |
+
"sae_top_50_test_accuracy": null,
|
71 |
+
"sae_top_100_test_accuracy": null
|
72 |
+
},
|
73 |
+
{
|
74 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
75 |
+
"llm_test_accuracy": 0.9544000387191772,
|
76 |
+
"llm_top_1_test_accuracy": 0.6764,
|
77 |
+
"llm_top_2_test_accuracy": 0.7150000000000001,
|
78 |
+
"llm_top_5_test_accuracy": 0.7592000000000001,
|
79 |
+
"llm_top_10_test_accuracy": null,
|
80 |
+
"llm_top_20_test_accuracy": null,
|
81 |
+
"llm_top_50_test_accuracy": null,
|
82 |
+
"llm_top_100_test_accuracy": null,
|
83 |
+
"sae_test_accuracy": 0.9494000554084778,
|
84 |
+
"sae_top_1_test_accuracy": 0.7806,
|
85 |
+
"sae_top_2_test_accuracy": 0.7844,
|
86 |
+
"sae_top_5_test_accuracy": 0.8248,
|
87 |
+
"sae_top_10_test_accuracy": null,
|
88 |
+
"sae_top_20_test_accuracy": null,
|
89 |
+
"sae_top_50_test_accuracy": null,
|
90 |
+
"sae_top_100_test_accuracy": null
|
91 |
+
},
|
92 |
+
{
|
93 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
94 |
+
"llm_test_accuracy": 0.9332000494003296,
|
95 |
+
"llm_top_1_test_accuracy": 0.6884,
|
96 |
+
"llm_top_2_test_accuracy": 0.737,
|
97 |
+
"llm_top_5_test_accuracy": 0.7492,
|
98 |
+
"llm_top_10_test_accuracy": null,
|
99 |
+
"llm_top_20_test_accuracy": null,
|
100 |
+
"llm_top_50_test_accuracy": null,
|
101 |
+
"llm_top_100_test_accuracy": null,
|
102 |
+
"sae_test_accuracy": 0.9282000422477722,
|
103 |
+
"sae_top_1_test_accuracy": 0.7826000000000001,
|
104 |
+
"sae_top_2_test_accuracy": 0.7958000000000001,
|
105 |
+
"sae_top_5_test_accuracy": 0.8308,
|
106 |
+
"sae_top_10_test_accuracy": null,
|
107 |
+
"sae_top_20_test_accuracy": null,
|
108 |
+
"sae_top_50_test_accuracy": null,
|
109 |
+
"sae_top_100_test_accuracy": null
|
110 |
+
},
|
111 |
+
{
|
112 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
113 |
+
"llm_test_accuracy": 0.9164000511169433,
|
114 |
+
"llm_top_1_test_accuracy": 0.598,
|
115 |
+
"llm_top_2_test_accuracy": 0.6406,
|
116 |
+
"llm_top_5_test_accuracy": 0.6748000000000001,
|
117 |
+
"llm_top_10_test_accuracy": null,
|
118 |
+
"llm_top_20_test_accuracy": null,
|
119 |
+
"llm_top_50_test_accuracy": null,
|
120 |
+
"llm_top_100_test_accuracy": null,
|
121 |
+
"sae_test_accuracy": 0.9142000317573548,
|
122 |
+
"sae_top_1_test_accuracy": 0.728,
|
123 |
+
"sae_top_2_test_accuracy": 0.7612,
|
124 |
+
"sae_top_5_test_accuracy": 0.8202,
|
125 |
+
"sae_top_10_test_accuracy": null,
|
126 |
+
"sae_top_20_test_accuracy": null,
|
127 |
+
"sae_top_50_test_accuracy": null,
|
128 |
+
"sae_top_100_test_accuracy": null
|
129 |
+
},
|
130 |
+
{
|
131 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
132 |
+
"llm_test_accuracy": 0.9820000529289246,
|
133 |
+
"llm_top_1_test_accuracy": 0.672,
|
134 |
+
"llm_top_2_test_accuracy": 0.724,
|
135 |
+
"llm_top_5_test_accuracy": 0.766,
|
136 |
+
"llm_top_10_test_accuracy": null,
|
137 |
+
"llm_top_20_test_accuracy": null,
|
138 |
+
"llm_top_50_test_accuracy": null,
|
139 |
+
"llm_top_100_test_accuracy": null,
|
140 |
+
"sae_test_accuracy": 0.9735000431537628,
|
141 |
+
"sae_top_1_test_accuracy": 0.909,
|
142 |
+
"sae_top_2_test_accuracy": 0.914,
|
143 |
+
"sae_top_5_test_accuracy": 0.953,
|
144 |
+
"sae_top_10_test_accuracy": null,
|
145 |
+
"sae_top_20_test_accuracy": null,
|
146 |
+
"sae_top_50_test_accuracy": null,
|
147 |
+
"sae_top_100_test_accuracy": null
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"dataset_name": "codeparrot/github-code_results",
|
151 |
+
"llm_test_accuracy": 0.9672000408172607,
|
152 |
+
"llm_top_1_test_accuracy": 0.6668000000000001,
|
153 |
+
"llm_top_2_test_accuracy": 0.6896,
|
154 |
+
"llm_top_5_test_accuracy": 0.7656000000000001,
|
155 |
+
"llm_top_10_test_accuracy": null,
|
156 |
+
"llm_top_20_test_accuracy": null,
|
157 |
+
"llm_top_50_test_accuracy": null,
|
158 |
+
"llm_top_100_test_accuracy": null,
|
159 |
+
"sae_test_accuracy": 0.9658000349998475,
|
160 |
+
"sae_top_1_test_accuracy": 0.6312,
|
161 |
+
"sae_top_2_test_accuracy": 0.7244000000000002,
|
162 |
+
"sae_top_5_test_accuracy": 0.841,
|
163 |
+
"sae_top_10_test_accuracy": null,
|
164 |
+
"sae_top_20_test_accuracy": null,
|
165 |
+
"sae_top_50_test_accuracy": null,
|
166 |
+
"sae_top_100_test_accuracy": null
|
167 |
+
},
|
168 |
+
{
|
169 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
170 |
+
"llm_test_accuracy": 0.9507500380277634,
|
171 |
+
"llm_top_1_test_accuracy": 0.63625,
|
172 |
+
"llm_top_2_test_accuracy": 0.7655000000000001,
|
173 |
+
"llm_top_5_test_accuracy": 0.82475,
|
174 |
+
"llm_top_10_test_accuracy": null,
|
175 |
+
"llm_top_20_test_accuracy": null,
|
176 |
+
"llm_top_50_test_accuracy": null,
|
177 |
+
"llm_top_100_test_accuracy": null,
|
178 |
+
"sae_test_accuracy": 0.9487500637769699,
|
179 |
+
"sae_top_1_test_accuracy": 0.7792499999999999,
|
180 |
+
"sae_top_2_test_accuracy": 0.893,
|
181 |
+
"sae_top_5_test_accuracy": 0.90125,
|
182 |
+
"sae_top_10_test_accuracy": null,
|
183 |
+
"sae_top_20_test_accuracy": null,
|
184 |
+
"sae_top_50_test_accuracy": null,
|
185 |
+
"sae_top_100_test_accuracy": null
|
186 |
+
},
|
187 |
+
{
|
188 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
189 |
+
"llm_test_accuracy": 0.9994000434875489,
|
190 |
+
"llm_top_1_test_accuracy": 0.737,
|
191 |
+
"llm_top_2_test_accuracy": 0.7796,
|
192 |
+
"llm_top_5_test_accuracy": 0.9126,
|
193 |
+
"llm_top_10_test_accuracy": null,
|
194 |
+
"llm_top_20_test_accuracy": null,
|
195 |
+
"llm_top_50_test_accuracy": null,
|
196 |
+
"llm_top_100_test_accuracy": null,
|
197 |
+
"sae_test_accuracy": 0.9992000579833984,
|
198 |
+
"sae_top_1_test_accuracy": 0.893,
|
199 |
+
"sae_top_2_test_accuracy": 0.9119999999999999,
|
200 |
+
"sae_top_5_test_accuracy": 0.9568,
|
201 |
+
"sae_top_10_test_accuracy": null,
|
202 |
+
"sae_top_20_test_accuracy": null,
|
203 |
+
"sae_top_50_test_accuracy": null,
|
204 |
+
"sae_top_100_test_accuracy": null
|
205 |
+
}
|
206 |
+
],
|
207 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
208 |
+
"sae_lens_id": "custom_sae",
|
209 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3",
|
210 |
+
"sae_lens_version": "5.3.2",
|
211 |
+
"sae_cfg_dict": {
|
212 |
+
"model_name": "gemma-2-2b",
|
213 |
+
"d_in": 2304,
|
214 |
+
"d_sae": 16384,
|
215 |
+
"hook_layer": 12,
|
216 |
+
"hook_name": "blocks.12.hook_resid_post",
|
217 |
+
"context_size": null,
|
218 |
+
"hook_head_index": null,
|
219 |
+
"architecture": "matryoshka_batch_topk",
|
220 |
+
"apply_b_dec_to_input": null,
|
221 |
+
"finetuning_scaling_factor": null,
|
222 |
+
"activation_fn_str": "",
|
223 |
+
"prepend_bos": true,
|
224 |
+
"normalize_activations": "none",
|
225 |
+
"dtype": "bfloat16",
|
226 |
+
"device": "",
|
227 |
+
"dataset_path": "",
|
228 |
+
"dataset_trust_remote_code": true,
|
229 |
+
"seqpos_slice": [
|
230 |
+
null
|
231 |
+
],
|
232 |
+
"training_tokens": -100000,
|
233 |
+
"sae_lens_training_version": null,
|
234 |
+
"neuronpedia_id": null
|
235 |
+
},
|
236 |
+
"eval_result_unstructured": {
|
237 |
+
"LabHC/bias_in_bios_class_set1_results": {
|
238 |
+
"sae_test_accuracy": {
|
239 |
+
"0": 0.9490000605583191,
|
240 |
+
"1": 0.9660000205039978,
|
241 |
+
"2": 0.9500000476837158,
|
242 |
+
"6": 0.9880000352859497,
|
243 |
+
"9": 0.9790000319480896
|
244 |
+
},
|
245 |
+
"llm_test_accuracy": {
|
246 |
+
"0": 0.9510000348091125,
|
247 |
+
"1": 0.9670000672340393,
|
248 |
+
"2": 0.9520000219345093,
|
249 |
+
"6": 0.9930000305175781,
|
250 |
+
"9": 0.984000027179718
|
251 |
+
},
|
252 |
+
"llm_top_1_test_accuracy": {
|
253 |
+
"0": 0.568,
|
254 |
+
"1": 0.629,
|
255 |
+
"2": 0.679,
|
256 |
+
"6": 0.791,
|
257 |
+
"9": 0.551
|
258 |
+
},
|
259 |
+
"llm_top_2_test_accuracy": {
|
260 |
+
"0": 0.585,
|
261 |
+
"1": 0.666,
|
262 |
+
"2": 0.673,
|
263 |
+
"6": 0.801,
|
264 |
+
"9": 0.712
|
265 |
+
},
|
266 |
+
"llm_top_5_test_accuracy": {
|
267 |
+
"0": 0.72,
|
268 |
+
"1": 0.707,
|
269 |
+
"2": 0.764,
|
270 |
+
"6": 0.899,
|
271 |
+
"9": 0.864
|
272 |
+
},
|
273 |
+
"sae_top_1_test_accuracy": {
|
274 |
+
"0": 0.589,
|
275 |
+
"1": 0.609,
|
276 |
+
"2": 0.691,
|
277 |
+
"6": 0.775,
|
278 |
+
"9": 0.944
|
279 |
+
},
|
280 |
+
"sae_top_2_test_accuracy": {
|
281 |
+
"0": 0.634,
|
282 |
+
"1": 0.618,
|
283 |
+
"2": 0.744,
|
284 |
+
"6": 0.978,
|
285 |
+
"9": 0.935
|
286 |
+
},
|
287 |
+
"sae_top_5_test_accuracy": {
|
288 |
+
"0": 0.868,
|
289 |
+
"1": 0.689,
|
290 |
+
"2": 0.902,
|
291 |
+
"6": 0.977,
|
292 |
+
"9": 0.944
|
293 |
+
}
|
294 |
+
},
|
295 |
+
"LabHC/bias_in_bios_class_set2_results": {
|
296 |
+
"sae_test_accuracy": {
|
297 |
+
"11": 0.956000030040741,
|
298 |
+
"13": 0.9530000686645508,
|
299 |
+
"14": 0.9530000686645508,
|
300 |
+
"18": 0.9300000667572021,
|
301 |
+
"19": 0.9550000429153442
|
302 |
+
},
|
303 |
+
"llm_test_accuracy": {
|
304 |
+
"11": 0.9650000333786011,
|
305 |
+
"13": 0.9520000219345093,
|
306 |
+
"14": 0.9550000429153442,
|
307 |
+
"18": 0.9360000491142273,
|
308 |
+
"19": 0.9640000462532043
|
309 |
+
},
|
310 |
+
"llm_top_1_test_accuracy": {
|
311 |
+
"11": 0.558,
|
312 |
+
"13": 0.673,
|
313 |
+
"14": 0.656,
|
314 |
+
"18": 0.702,
|
315 |
+
"19": 0.793
|
316 |
+
},
|
317 |
+
"llm_top_2_test_accuracy": {
|
318 |
+
"11": 0.686,
|
319 |
+
"13": 0.713,
|
320 |
+
"14": 0.687,
|
321 |
+
"18": 0.724,
|
322 |
+
"19": 0.765
|
323 |
+
},
|
324 |
+
"llm_top_5_test_accuracy": {
|
325 |
+
"11": 0.782,
|
326 |
+
"13": 0.742,
|
327 |
+
"14": 0.716,
|
328 |
+
"18": 0.725,
|
329 |
+
"19": 0.831
|
330 |
+
},
|
331 |
+
"sae_top_1_test_accuracy": {
|
332 |
+
"11": 0.835,
|
333 |
+
"13": 0.668,
|
334 |
+
"14": 0.854,
|
335 |
+
"18": 0.706,
|
336 |
+
"19": 0.84
|
337 |
+
},
|
338 |
+
"sae_top_2_test_accuracy": {
|
339 |
+
"11": 0.854,
|
340 |
+
"13": 0.674,
|
341 |
+
"14": 0.856,
|
342 |
+
"18": 0.694,
|
343 |
+
"19": 0.844
|
344 |
+
},
|
345 |
+
"sae_top_5_test_accuracy": {
|
346 |
+
"11": 0.848,
|
347 |
+
"13": 0.805,
|
348 |
+
"14": 0.876,
|
349 |
+
"18": 0.742,
|
350 |
+
"19": 0.853
|
351 |
+
}
|
352 |
+
},
|
353 |
+
"LabHC/bias_in_bios_class_set3_results": {
|
354 |
+
"sae_test_accuracy": {
|
355 |
+
"20": 0.9610000252723694,
|
356 |
+
"21": 0.9190000295639038,
|
357 |
+
"22": 0.9170000553131104,
|
358 |
+
"25": 0.9620000720024109,
|
359 |
+
"26": 0.8820000290870667
|
360 |
+
},
|
361 |
+
"llm_test_accuracy": {
|
362 |
+
"20": 0.9610000252723694,
|
363 |
+
"21": 0.9270000457763672,
|
364 |
+
"22": 0.921000063419342,
|
365 |
+
"25": 0.968000054359436,
|
366 |
+
"26": 0.8890000581741333
|
367 |
+
},
|
368 |
+
"llm_top_1_test_accuracy": {
|
369 |
+
"20": 0.713,
|
370 |
+
"21": 0.766,
|
371 |
+
"22": 0.639,
|
372 |
+
"25": 0.69,
|
373 |
+
"26": 0.634
|
374 |
+
},
|
375 |
+
"llm_top_2_test_accuracy": {
|
376 |
+
"20": 0.804,
|
377 |
+
"21": 0.777,
|
378 |
+
"22": 0.68,
|
379 |
+
"25": 0.759,
|
380 |
+
"26": 0.665
|
381 |
+
},
|
382 |
+
"llm_top_5_test_accuracy": {
|
383 |
+
"20": 0.813,
|
384 |
+
"21": 0.777,
|
385 |
+
"22": 0.686,
|
386 |
+
"25": 0.798,
|
387 |
+
"26": 0.672
|
388 |
+
},
|
389 |
+
"sae_top_1_test_accuracy": {
|
390 |
+
"20": 0.869,
|
391 |
+
"21": 0.769,
|
392 |
+
"22": 0.764,
|
393 |
+
"25": 0.87,
|
394 |
+
"26": 0.641
|
395 |
+
},
|
396 |
+
"sae_top_2_test_accuracy": {
|
397 |
+
"20": 0.906,
|
398 |
+
"21": 0.777,
|
399 |
+
"22": 0.759,
|
400 |
+
"25": 0.858,
|
401 |
+
"26": 0.679
|
402 |
+
},
|
403 |
+
"sae_top_5_test_accuracy": {
|
404 |
+
"20": 0.922,
|
405 |
+
"21": 0.819,
|
406 |
+
"22": 0.794,
|
407 |
+
"25": 0.859,
|
408 |
+
"26": 0.76
|
409 |
+
}
|
410 |
+
},
|
411 |
+
"canrager/amazon_reviews_mcauley_1and5_results": {
|
412 |
+
"sae_test_accuracy": {
|
413 |
+
"1": 0.9430000185966492,
|
414 |
+
"2": 0.9330000281333923,
|
415 |
+
"3": 0.9240000247955322,
|
416 |
+
"5": 0.9130000472068787,
|
417 |
+
"6": 0.8580000400543213
|
418 |
+
},
|
419 |
+
"llm_test_accuracy": {
|
420 |
+
"1": 0.9480000734329224,
|
421 |
+
"2": 0.9350000619888306,
|
422 |
+
"3": 0.9140000343322754,
|
423 |
+
"5": 0.9110000729560852,
|
424 |
+
"6": 0.8740000128746033
|
425 |
+
},
|
426 |
+
"llm_top_1_test_accuracy": {
|
427 |
+
"1": 0.668,
|
428 |
+
"2": 0.573,
|
429 |
+
"3": 0.591,
|
430 |
+
"5": 0.568,
|
431 |
+
"6": 0.59
|
432 |
+
},
|
433 |
+
"llm_top_2_test_accuracy": {
|
434 |
+
"1": 0.745,
|
435 |
+
"2": 0.636,
|
436 |
+
"3": 0.606,
|
437 |
+
"5": 0.595,
|
438 |
+
"6": 0.621
|
439 |
+
},
|
440 |
+
"llm_top_5_test_accuracy": {
|
441 |
+
"1": 0.778,
|
442 |
+
"2": 0.643,
|
443 |
+
"3": 0.608,
|
444 |
+
"5": 0.66,
|
445 |
+
"6": 0.685
|
446 |
+
},
|
447 |
+
"sae_top_1_test_accuracy": {
|
448 |
+
"1": 0.827,
|
449 |
+
"2": 0.64,
|
450 |
+
"3": 0.561,
|
451 |
+
"5": 0.871,
|
452 |
+
"6": 0.741
|
453 |
+
},
|
454 |
+
"sae_top_2_test_accuracy": {
|
455 |
+
"1": 0.849,
|
456 |
+
"2": 0.663,
|
457 |
+
"3": 0.681,
|
458 |
+
"5": 0.859,
|
459 |
+
"6": 0.754
|
460 |
+
},
|
461 |
+
"sae_top_5_test_accuracy": {
|
462 |
+
"1": 0.868,
|
463 |
+
"2": 0.889,
|
464 |
+
"3": 0.706,
|
465 |
+
"5": 0.867,
|
466 |
+
"6": 0.771
|
467 |
+
}
|
468 |
+
},
|
469 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment_results": {
|
470 |
+
"sae_test_accuracy": {
|
471 |
+
"1.0": 0.9730000495910645,
|
472 |
+
"5.0": 0.9740000367164612
|
473 |
+
},
|
474 |
+
"llm_test_accuracy": {
|
475 |
+
"1.0": 0.9810000658035278,
|
476 |
+
"5.0": 0.9830000400543213
|
477 |
+
},
|
478 |
+
"llm_top_1_test_accuracy": {
|
479 |
+
"1.0": 0.672,
|
480 |
+
"5.0": 0.672
|
481 |
+
},
|
482 |
+
"llm_top_2_test_accuracy": {
|
483 |
+
"1.0": 0.724,
|
484 |
+
"5.0": 0.724
|
485 |
+
},
|
486 |
+
"llm_top_5_test_accuracy": {
|
487 |
+
"1.0": 0.766,
|
488 |
+
"5.0": 0.766
|
489 |
+
},
|
490 |
+
"sae_top_1_test_accuracy": {
|
491 |
+
"1.0": 0.909,
|
492 |
+
"5.0": 0.909
|
493 |
+
},
|
494 |
+
"sae_top_2_test_accuracy": {
|
495 |
+
"1.0": 0.914,
|
496 |
+
"5.0": 0.914
|
497 |
+
},
|
498 |
+
"sae_top_5_test_accuracy": {
|
499 |
+
"1.0": 0.953,
|
500 |
+
"5.0": 0.953
|
501 |
+
}
|
502 |
+
},
|
503 |
+
"codeparrot/github-code_results": {
|
504 |
+
"sae_test_accuracy": {
|
505 |
+
"C": 0.9570000171661377,
|
506 |
+
"Python": 0.9810000658035278,
|
507 |
+
"HTML": 0.984000027179718,
|
508 |
+
"Java": 0.9550000429153442,
|
509 |
+
"PHP": 0.9520000219345093
|
510 |
+
},
|
511 |
+
"llm_test_accuracy": {
|
512 |
+
"C": 0.9510000348091125,
|
513 |
+
"Python": 0.987000048160553,
|
514 |
+
"HTML": 0.984000027179718,
|
515 |
+
"Java": 0.968000054359436,
|
516 |
+
"PHP": 0.9460000395774841
|
517 |
+
},
|
518 |
+
"llm_top_1_test_accuracy": {
|
519 |
+
"C": 0.655,
|
520 |
+
"Python": 0.661,
|
521 |
+
"HTML": 0.802,
|
522 |
+
"Java": 0.622,
|
523 |
+
"PHP": 0.594
|
524 |
+
},
|
525 |
+
"llm_top_2_test_accuracy": {
|
526 |
+
"C": 0.667,
|
527 |
+
"Python": 0.668,
|
528 |
+
"HTML": 0.798,
|
529 |
+
"Java": 0.674,
|
530 |
+
"PHP": 0.641
|
531 |
+
},
|
532 |
+
"llm_top_5_test_accuracy": {
|
533 |
+
"C": 0.78,
|
534 |
+
"Python": 0.73,
|
535 |
+
"HTML": 0.906,
|
536 |
+
"Java": 0.728,
|
537 |
+
"PHP": 0.684
|
538 |
+
},
|
539 |
+
"sae_top_1_test_accuracy": {
|
540 |
+
"C": 0.638,
|
541 |
+
"Python": 0.609,
|
542 |
+
"HTML": 0.706,
|
543 |
+
"Java": 0.594,
|
544 |
+
"PHP": 0.609
|
545 |
+
},
|
546 |
+
"sae_top_2_test_accuracy": {
|
547 |
+
"C": 0.625,
|
548 |
+
"Python": 0.643,
|
549 |
+
"HTML": 0.935,
|
550 |
+
"Java": 0.655,
|
551 |
+
"PHP": 0.764
|
552 |
+
},
|
553 |
+
"sae_top_5_test_accuracy": {
|
554 |
+
"C": 0.883,
|
555 |
+
"Python": 0.729,
|
556 |
+
"HTML": 0.959,
|
557 |
+
"Java": 0.705,
|
558 |
+
"PHP": 0.929
|
559 |
+
}
|
560 |
+
},
|
561 |
+
"fancyzhx/ag_news_results": {
|
562 |
+
"sae_test_accuracy": {
|
563 |
+
"0": 0.9390000700950623,
|
564 |
+
"1": 0.9820000529289246,
|
565 |
+
"2": 0.9300000667572021,
|
566 |
+
"3": 0.9440000653266907
|
567 |
+
},
|
568 |
+
"llm_test_accuracy": {
|
569 |
+
"0": 0.9360000491142273,
|
570 |
+
"1": 0.9890000224113464,
|
571 |
+
"2": 0.9280000329017639,
|
572 |
+
"3": 0.9500000476837158
|
573 |
+
},
|
574 |
+
"llm_top_1_test_accuracy": {
|
575 |
+
"0": 0.58,
|
576 |
+
"1": 0.676,
|
577 |
+
"2": 0.667,
|
578 |
+
"3": 0.622
|
579 |
+
},
|
580 |
+
"llm_top_2_test_accuracy": {
|
581 |
+
"0": 0.811,
|
582 |
+
"1": 0.798,
|
583 |
+
"2": 0.693,
|
584 |
+
"3": 0.76
|
585 |
+
},
|
586 |
+
"llm_top_5_test_accuracy": {
|
587 |
+
"0": 0.815,
|
588 |
+
"1": 0.884,
|
589 |
+
"2": 0.756,
|
590 |
+
"3": 0.844
|
591 |
+
},
|
592 |
+
"sae_top_1_test_accuracy": {
|
593 |
+
"0": 0.828,
|
594 |
+
"1": 0.972,
|
595 |
+
"2": 0.659,
|
596 |
+
"3": 0.658
|
597 |
+
},
|
598 |
+
"sae_top_2_test_accuracy": {
|
599 |
+
"0": 0.896,
|
600 |
+
"1": 0.977,
|
601 |
+
"2": 0.865,
|
602 |
+
"3": 0.834
|
603 |
+
},
|
604 |
+
"sae_top_5_test_accuracy": {
|
605 |
+
"0": 0.891,
|
606 |
+
"1": 0.978,
|
607 |
+
"2": 0.862,
|
608 |
+
"3": 0.874
|
609 |
+
}
|
610 |
+
},
|
611 |
+
"Helsinki-NLP/europarl_results": {
|
612 |
+
"sae_test_accuracy": {
|
613 |
+
"en": 0.999000072479248,
|
614 |
+
"fr": 1.0,
|
615 |
+
"de": 0.999000072479248,
|
616 |
+
"es": 0.999000072479248,
|
617 |
+
"nl": 0.999000072479248
|
618 |
+
},
|
619 |
+
"llm_test_accuracy": {
|
620 |
+
"en": 1.0,
|
621 |
+
"fr": 0.999000072479248,
|
622 |
+
"de": 0.999000072479248,
|
623 |
+
"es": 1.0,
|
624 |
+
"nl": 0.999000072479248
|
625 |
+
},
|
626 |
+
"llm_top_1_test_accuracy": {
|
627 |
+
"en": 0.748,
|
628 |
+
"fr": 0.615,
|
629 |
+
"de": 0.761,
|
630 |
+
"es": 0.904,
|
631 |
+
"nl": 0.657
|
632 |
+
},
|
633 |
+
"llm_top_2_test_accuracy": {
|
634 |
+
"en": 0.829,
|
635 |
+
"fr": 0.583,
|
636 |
+
"de": 0.829,
|
637 |
+
"es": 0.914,
|
638 |
+
"nl": 0.743
|
639 |
+
},
|
640 |
+
"llm_top_5_test_accuracy": {
|
641 |
+
"en": 0.892,
|
642 |
+
"fr": 0.914,
|
643 |
+
"de": 0.917,
|
644 |
+
"es": 0.984,
|
645 |
+
"nl": 0.856
|
646 |
+
},
|
647 |
+
"sae_top_1_test_accuracy": {
|
648 |
+
"en": 0.999,
|
649 |
+
"fr": 0.974,
|
650 |
+
"de": 0.899,
|
651 |
+
"es": 0.967,
|
652 |
+
"nl": 0.626
|
653 |
+
},
|
654 |
+
"sae_top_2_test_accuracy": {
|
655 |
+
"en": 0.999,
|
656 |
+
"fr": 0.983,
|
657 |
+
"de": 0.906,
|
658 |
+
"es": 0.968,
|
659 |
+
"nl": 0.704
|
660 |
+
},
|
661 |
+
"sae_top_5_test_accuracy": {
|
662 |
+
"en": 0.999,
|
663 |
+
"fr": 0.995,
|
664 |
+
"de": 0.951,
|
665 |
+
"es": 0.969,
|
666 |
+
"nl": 0.87
|
667 |
+
}
|
668 |
+
}
|
669 |
+
}
|
670 |
+
}
|
sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,670 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "sparse_probing",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"LabHC/bias_in_bios_class_set2",
|
8 |
+
"LabHC/bias_in_bios_class_set3",
|
9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
11 |
+
"codeparrot/github-code",
|
12 |
+
"fancyzhx/ag_news",
|
13 |
+
"Helsinki-NLP/europarl"
|
14 |
+
],
|
15 |
+
"probe_train_set_size": 4000,
|
16 |
+
"probe_test_set_size": 1000,
|
17 |
+
"context_length": 128,
|
18 |
+
"sae_batch_size": 125,
|
19 |
+
"llm_batch_size": 32,
|
20 |
+
"llm_dtype": "bfloat16",
|
21 |
+
"model_name": "gemma-2-2b",
|
22 |
+
"k_values": [
|
23 |
+
1,
|
24 |
+
2,
|
25 |
+
5
|
26 |
+
],
|
27 |
+
"lower_vram_usage": false
|
28 |
+
},
|
29 |
+
"eval_id": "5c2e1585-6b87-48c3-b542-418db3bbcce6",
|
30 |
+
"datetime_epoch_millis": 1737823966669,
|
31 |
+
"eval_result_metrics": {
|
32 |
+
"llm": {
|
33 |
+
"llm_test_accuracy": 0.9590937938541175,
|
34 |
+
"llm_top_1_test_accuracy": 0.66480625,
|
35 |
+
"llm_top_2_test_accuracy": 0.7173375000000001,
|
36 |
+
"llm_top_5_test_accuracy": 0.78036875,
|
37 |
+
"llm_top_10_test_accuracy": null,
|
38 |
+
"llm_top_20_test_accuracy": null,
|
39 |
+
"llm_top_50_test_accuracy": null,
|
40 |
+
"llm_top_100_test_accuracy": null
|
41 |
+
},
|
42 |
+
"sae": {
|
43 |
+
"sae_test_accuracy": 0.9580750439316035,
|
44 |
+
"sae_top_1_test_accuracy": 0.7433750000000001,
|
45 |
+
"sae_top_2_test_accuracy": 0.81263125,
|
46 |
+
"sae_top_5_test_accuracy": 0.8634062499999999,
|
47 |
+
"sae_top_10_test_accuracy": null,
|
48 |
+
"sae_top_20_test_accuracy": null,
|
49 |
+
"sae_top_50_test_accuracy": null,
|
50 |
+
"sae_top_100_test_accuracy": null
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"eval_result_details": [
|
54 |
+
{
|
55 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
56 |
+
"llm_test_accuracy": 0.9694000363349915,
|
57 |
+
"llm_top_1_test_accuracy": 0.6436000000000001,
|
58 |
+
"llm_top_2_test_accuracy": 0.6874,
|
59 |
+
"llm_top_5_test_accuracy": 0.7908,
|
60 |
+
"llm_top_10_test_accuracy": null,
|
61 |
+
"llm_top_20_test_accuracy": null,
|
62 |
+
"llm_top_50_test_accuracy": null,
|
63 |
+
"llm_top_100_test_accuracy": null,
|
64 |
+
"sae_test_accuracy": 0.9702000498771668,
|
65 |
+
"sae_top_1_test_accuracy": 0.7598,
|
66 |
+
"sae_top_2_test_accuracy": 0.7984,
|
67 |
+
"sae_top_5_test_accuracy": 0.8625999999999999,
|
68 |
+
"sae_top_10_test_accuracy": null,
|
69 |
+
"sae_top_20_test_accuracy": null,
|
70 |
+
"sae_top_50_test_accuracy": null,
|
71 |
+
"sae_top_100_test_accuracy": null
|
72 |
+
},
|
73 |
+
{
|
74 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
75 |
+
"llm_test_accuracy": 0.9544000387191772,
|
76 |
+
"llm_top_1_test_accuracy": 0.6764,
|
77 |
+
"llm_top_2_test_accuracy": 0.7150000000000001,
|
78 |
+
"llm_top_5_test_accuracy": 0.7592000000000001,
|
79 |
+
"llm_top_10_test_accuracy": null,
|
80 |
+
"llm_top_20_test_accuracy": null,
|
81 |
+
"llm_top_50_test_accuracy": null,
|
82 |
+
"llm_top_100_test_accuracy": null,
|
83 |
+
"sae_test_accuracy": 0.9516000390052796,
|
84 |
+
"sae_top_1_test_accuracy": 0.7436,
|
85 |
+
"sae_top_2_test_accuracy": 0.764,
|
86 |
+
"sae_top_5_test_accuracy": 0.8188000000000001,
|
87 |
+
"sae_top_10_test_accuracy": null,
|
88 |
+
"sae_top_20_test_accuracy": null,
|
89 |
+
"sae_top_50_test_accuracy": null,
|
90 |
+
"sae_top_100_test_accuracy": null
|
91 |
+
},
|
92 |
+
{
|
93 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
94 |
+
"llm_test_accuracy": 0.9332000494003296,
|
95 |
+
"llm_top_1_test_accuracy": 0.6884,
|
96 |
+
"llm_top_2_test_accuracy": 0.737,
|
97 |
+
"llm_top_5_test_accuracy": 0.7492,
|
98 |
+
"llm_top_10_test_accuracy": null,
|
99 |
+
"llm_top_20_test_accuracy": null,
|
100 |
+
"llm_top_50_test_accuracy": null,
|
101 |
+
"llm_top_100_test_accuracy": null,
|
102 |
+
"sae_test_accuracy": 0.9320000648498535,
|
103 |
+
"sae_top_1_test_accuracy": 0.7198,
|
104 |
+
"sae_top_2_test_accuracy": 0.7842,
|
105 |
+
"sae_top_5_test_accuracy": 0.8291999999999999,
|
106 |
+
"sae_top_10_test_accuracy": null,
|
107 |
+
"sae_top_20_test_accuracy": null,
|
108 |
+
"sae_top_50_test_accuracy": null,
|
109 |
+
"sae_top_100_test_accuracy": null
|
110 |
+
},
|
111 |
+
{
|
112 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
113 |
+
"llm_test_accuracy": 0.9164000511169433,
|
114 |
+
"llm_top_1_test_accuracy": 0.598,
|
115 |
+
"llm_top_2_test_accuracy": 0.6406,
|
116 |
+
"llm_top_5_test_accuracy": 0.6748000000000001,
|
117 |
+
"llm_top_10_test_accuracy": null,
|
118 |
+
"llm_top_20_test_accuracy": null,
|
119 |
+
"llm_top_50_test_accuracy": null,
|
120 |
+
"llm_top_100_test_accuracy": null,
|
121 |
+
"sae_test_accuracy": 0.9206000447273255,
|
122 |
+
"sae_top_1_test_accuracy": 0.715,
|
123 |
+
"sae_top_2_test_accuracy": 0.7906000000000001,
|
124 |
+
"sae_top_5_test_accuracy": 0.8022,
|
125 |
+
"sae_top_10_test_accuracy": null,
|
126 |
+
"sae_top_20_test_accuracy": null,
|
127 |
+
"sae_top_50_test_accuracy": null,
|
128 |
+
"sae_top_100_test_accuracy": null
|
129 |
+
},
|
130 |
+
{
|
131 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
132 |
+
"llm_test_accuracy": 0.9820000529289246,
|
133 |
+
"llm_top_1_test_accuracy": 0.672,
|
134 |
+
"llm_top_2_test_accuracy": 0.724,
|
135 |
+
"llm_top_5_test_accuracy": 0.766,
|
136 |
+
"llm_top_10_test_accuracy": null,
|
137 |
+
"llm_top_20_test_accuracy": null,
|
138 |
+
"llm_top_50_test_accuracy": null,
|
139 |
+
"llm_top_100_test_accuracy": null,
|
140 |
+
"sae_test_accuracy": 0.9730000495910645,
|
141 |
+
"sae_top_1_test_accuracy": 0.714,
|
142 |
+
"sae_top_2_test_accuracy": 0.857,
|
143 |
+
"sae_top_5_test_accuracy": 0.882,
|
144 |
+
"sae_top_10_test_accuracy": null,
|
145 |
+
"sae_top_20_test_accuracy": null,
|
146 |
+
"sae_top_50_test_accuracy": null,
|
147 |
+
"sae_top_100_test_accuracy": null
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"dataset_name": "codeparrot/github-code_results",
|
151 |
+
"llm_test_accuracy": 0.9672000408172607,
|
152 |
+
"llm_top_1_test_accuracy": 0.6668000000000001,
|
153 |
+
"llm_top_2_test_accuracy": 0.6896,
|
154 |
+
"llm_top_5_test_accuracy": 0.7656000000000001,
|
155 |
+
"llm_top_10_test_accuracy": null,
|
156 |
+
"llm_top_20_test_accuracy": null,
|
157 |
+
"llm_top_50_test_accuracy": null,
|
158 |
+
"llm_top_100_test_accuracy": null,
|
159 |
+
"sae_test_accuracy": 0.9670000433921814,
|
160 |
+
"sae_top_1_test_accuracy": 0.6804,
|
161 |
+
"sae_top_2_test_accuracy": 0.6994,
|
162 |
+
"sae_top_5_test_accuracy": 0.8422000000000001,
|
163 |
+
"sae_top_10_test_accuracy": null,
|
164 |
+
"sae_top_20_test_accuracy": null,
|
165 |
+
"sae_top_50_test_accuracy": null,
|
166 |
+
"sae_top_100_test_accuracy": null
|
167 |
+
},
|
168 |
+
{
|
169 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
170 |
+
"llm_test_accuracy": 0.9507500380277634,
|
171 |
+
"llm_top_1_test_accuracy": 0.63625,
|
172 |
+
"llm_top_2_test_accuracy": 0.7655000000000001,
|
173 |
+
"llm_top_5_test_accuracy": 0.82475,
|
174 |
+
"llm_top_10_test_accuracy": null,
|
175 |
+
"llm_top_20_test_accuracy": null,
|
176 |
+
"llm_top_50_test_accuracy": null,
|
177 |
+
"llm_top_100_test_accuracy": null,
|
178 |
+
"sae_test_accuracy": 0.9510000497102737,
|
179 |
+
"sae_top_1_test_accuracy": 0.77,
|
180 |
+
"sae_top_2_test_accuracy": 0.85625,
|
181 |
+
"sae_top_5_test_accuracy": 0.90325,
|
182 |
+
"sae_top_10_test_accuracy": null,
|
183 |
+
"sae_top_20_test_accuracy": null,
|
184 |
+
"sae_top_50_test_accuracy": null,
|
185 |
+
"sae_top_100_test_accuracy": null
|
186 |
+
},
|
187 |
+
{
|
188 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
189 |
+
"llm_test_accuracy": 0.9994000434875489,
|
190 |
+
"llm_top_1_test_accuracy": 0.737,
|
191 |
+
"llm_top_2_test_accuracy": 0.7796,
|
192 |
+
"llm_top_5_test_accuracy": 0.9126,
|
193 |
+
"llm_top_10_test_accuracy": null,
|
194 |
+
"llm_top_20_test_accuracy": null,
|
195 |
+
"llm_top_50_test_accuracy": null,
|
196 |
+
"llm_top_100_test_accuracy": null,
|
197 |
+
"sae_test_accuracy": 0.9992000102996826,
|
198 |
+
"sae_top_1_test_accuracy": 0.8444,
|
199 |
+
"sae_top_2_test_accuracy": 0.9512,
|
200 |
+
"sae_top_5_test_accuracy": 0.967,
|
201 |
+
"sae_top_10_test_accuracy": null,
|
202 |
+
"sae_top_20_test_accuracy": null,
|
203 |
+
"sae_top_50_test_accuracy": null,
|
204 |
+
"sae_top_100_test_accuracy": null
|
205 |
+
}
|
206 |
+
],
|
207 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
208 |
+
"sae_lens_id": "custom_sae",
|
209 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4",
|
210 |
+
"sae_lens_version": "5.3.2",
|
211 |
+
"sae_cfg_dict": {
|
212 |
+
"model_name": "gemma-2-2b",
|
213 |
+
"d_in": 2304,
|
214 |
+
"d_sae": 16384,
|
215 |
+
"hook_layer": 12,
|
216 |
+
"hook_name": "blocks.12.hook_resid_post",
|
217 |
+
"context_size": null,
|
218 |
+
"hook_head_index": null,
|
219 |
+
"architecture": "matryoshka_batch_topk",
|
220 |
+
"apply_b_dec_to_input": null,
|
221 |
+
"finetuning_scaling_factor": null,
|
222 |
+
"activation_fn_str": "",
|
223 |
+
"prepend_bos": true,
|
224 |
+
"normalize_activations": "none",
|
225 |
+
"dtype": "bfloat16",
|
226 |
+
"device": "",
|
227 |
+
"dataset_path": "",
|
228 |
+
"dataset_trust_remote_code": true,
|
229 |
+
"seqpos_slice": [
|
230 |
+
null
|
231 |
+
],
|
232 |
+
"training_tokens": -100000,
|
233 |
+
"sae_lens_training_version": null,
|
234 |
+
"neuronpedia_id": null
|
235 |
+
},
|
236 |
+
"eval_result_unstructured": {
|
237 |
+
"LabHC/bias_in_bios_class_set1_results": {
|
238 |
+
"sae_test_accuracy": {
|
239 |
+
"0": 0.9510000348091125,
|
240 |
+
"1": 0.9690000414848328,
|
241 |
+
"2": 0.9580000638961792,
|
242 |
+
"6": 0.9910000562667847,
|
243 |
+
"9": 0.9820000529289246
|
244 |
+
},
|
245 |
+
"llm_test_accuracy": {
|
246 |
+
"0": 0.9510000348091125,
|
247 |
+
"1": 0.9670000672340393,
|
248 |
+
"2": 0.9520000219345093,
|
249 |
+
"6": 0.9930000305175781,
|
250 |
+
"9": 0.984000027179718
|
251 |
+
},
|
252 |
+
"llm_top_1_test_accuracy": {
|
253 |
+
"0": 0.568,
|
254 |
+
"1": 0.629,
|
255 |
+
"2": 0.679,
|
256 |
+
"6": 0.791,
|
257 |
+
"9": 0.551
|
258 |
+
},
|
259 |
+
"llm_top_2_test_accuracy": {
|
260 |
+
"0": 0.585,
|
261 |
+
"1": 0.666,
|
262 |
+
"2": 0.673,
|
263 |
+
"6": 0.801,
|
264 |
+
"9": 0.712
|
265 |
+
},
|
266 |
+
"llm_top_5_test_accuracy": {
|
267 |
+
"0": 0.72,
|
268 |
+
"1": 0.707,
|
269 |
+
"2": 0.764,
|
270 |
+
"6": 0.899,
|
271 |
+
"9": 0.864
|
272 |
+
},
|
273 |
+
"sae_top_1_test_accuracy": {
|
274 |
+
"0": 0.573,
|
275 |
+
"1": 0.67,
|
276 |
+
"2": 0.636,
|
277 |
+
"6": 0.979,
|
278 |
+
"9": 0.941
|
279 |
+
},
|
280 |
+
"sae_top_2_test_accuracy": {
|
281 |
+
"0": 0.693,
|
282 |
+
"1": 0.673,
|
283 |
+
"2": 0.711,
|
284 |
+
"6": 0.979,
|
285 |
+
"9": 0.936
|
286 |
+
},
|
287 |
+
"sae_top_5_test_accuracy": {
|
288 |
+
"0": 0.743,
|
289 |
+
"1": 0.767,
|
290 |
+
"2": 0.873,
|
291 |
+
"6": 0.977,
|
292 |
+
"9": 0.953
|
293 |
+
}
|
294 |
+
},
|
295 |
+
"LabHC/bias_in_bios_class_set2_results": {
|
296 |
+
"sae_test_accuracy": {
|
297 |
+
"11": 0.9520000219345093,
|
298 |
+
"13": 0.9570000171661377,
|
299 |
+
"14": 0.9580000638961792,
|
300 |
+
"18": 0.9290000200271606,
|
301 |
+
"19": 0.9620000720024109
|
302 |
+
},
|
303 |
+
"llm_test_accuracy": {
|
304 |
+
"11": 0.9650000333786011,
|
305 |
+
"13": 0.9520000219345093,
|
306 |
+
"14": 0.9550000429153442,
|
307 |
+
"18": 0.9360000491142273,
|
308 |
+
"19": 0.9640000462532043
|
309 |
+
},
|
310 |
+
"llm_top_1_test_accuracy": {
|
311 |
+
"11": 0.558,
|
312 |
+
"13": 0.673,
|
313 |
+
"14": 0.656,
|
314 |
+
"18": 0.702,
|
315 |
+
"19": 0.793
|
316 |
+
},
|
317 |
+
"llm_top_2_test_accuracy": {
|
318 |
+
"11": 0.686,
|
319 |
+
"13": 0.713,
|
320 |
+
"14": 0.687,
|
321 |
+
"18": 0.724,
|
322 |
+
"19": 0.765
|
323 |
+
},
|
324 |
+
"llm_top_5_test_accuracy": {
|
325 |
+
"11": 0.782,
|
326 |
+
"13": 0.742,
|
327 |
+
"14": 0.716,
|
328 |
+
"18": 0.725,
|
329 |
+
"19": 0.831
|
330 |
+
},
|
331 |
+
"sae_top_1_test_accuracy": {
|
332 |
+
"11": 0.705,
|
333 |
+
"13": 0.664,
|
334 |
+
"14": 0.81,
|
335 |
+
"18": 0.69,
|
336 |
+
"19": 0.849
|
337 |
+
},
|
338 |
+
"sae_top_2_test_accuracy": {
|
339 |
+
"11": 0.726,
|
340 |
+
"13": 0.715,
|
341 |
+
"14": 0.805,
|
342 |
+
"18": 0.728,
|
343 |
+
"19": 0.846
|
344 |
+
},
|
345 |
+
"sae_top_5_test_accuracy": {
|
346 |
+
"11": 0.867,
|
347 |
+
"13": 0.738,
|
348 |
+
"14": 0.811,
|
349 |
+
"18": 0.81,
|
350 |
+
"19": 0.868
|
351 |
+
}
|
352 |
+
},
|
353 |
+
"LabHC/bias_in_bios_class_set3_results": {
|
354 |
+
"sae_test_accuracy": {
|
355 |
+
"20": 0.9630000591278076,
|
356 |
+
"21": 0.9250000715255737,
|
357 |
+
"22": 0.921000063419342,
|
358 |
+
"25": 0.9580000638961792,
|
359 |
+
"26": 0.893000066280365
|
360 |
+
},
|
361 |
+
"llm_test_accuracy": {
|
362 |
+
"20": 0.9610000252723694,
|
363 |
+
"21": 0.9270000457763672,
|
364 |
+
"22": 0.921000063419342,
|
365 |
+
"25": 0.968000054359436,
|
366 |
+
"26": 0.8890000581741333
|
367 |
+
},
|
368 |
+
"llm_top_1_test_accuracy": {
|
369 |
+
"20": 0.713,
|
370 |
+
"21": 0.766,
|
371 |
+
"22": 0.639,
|
372 |
+
"25": 0.69,
|
373 |
+
"26": 0.634
|
374 |
+
},
|
375 |
+
"llm_top_2_test_accuracy": {
|
376 |
+
"20": 0.804,
|
377 |
+
"21": 0.777,
|
378 |
+
"22": 0.68,
|
379 |
+
"25": 0.759,
|
380 |
+
"26": 0.665
|
381 |
+
},
|
382 |
+
"llm_top_5_test_accuracy": {
|
383 |
+
"20": 0.813,
|
384 |
+
"21": 0.777,
|
385 |
+
"22": 0.686,
|
386 |
+
"25": 0.798,
|
387 |
+
"26": 0.672
|
388 |
+
},
|
389 |
+
"sae_top_1_test_accuracy": {
|
390 |
+
"20": 0.677,
|
391 |
+
"21": 0.615,
|
392 |
+
"22": 0.749,
|
393 |
+
"25": 0.895,
|
394 |
+
"26": 0.663
|
395 |
+
},
|
396 |
+
"sae_top_2_test_accuracy": {
|
397 |
+
"20": 0.898,
|
398 |
+
"21": 0.713,
|
399 |
+
"22": 0.744,
|
400 |
+
"25": 0.888,
|
401 |
+
"26": 0.678
|
402 |
+
},
|
403 |
+
"sae_top_5_test_accuracy": {
|
404 |
+
"20": 0.916,
|
405 |
+
"21": 0.825,
|
406 |
+
"22": 0.794,
|
407 |
+
"25": 0.895,
|
408 |
+
"26": 0.716
|
409 |
+
}
|
410 |
+
},
|
411 |
+
"canrager/amazon_reviews_mcauley_1and5_results": {
|
412 |
+
"sae_test_accuracy": {
|
413 |
+
"1": 0.9480000734329224,
|
414 |
+
"2": 0.9280000329017639,
|
415 |
+
"3": 0.9270000457763672,
|
416 |
+
"5": 0.9340000152587891,
|
417 |
+
"6": 0.8660000562667847
|
418 |
+
},
|
419 |
+
"llm_test_accuracy": {
|
420 |
+
"1": 0.9480000734329224,
|
421 |
+
"2": 0.9350000619888306,
|
422 |
+
"3": 0.9140000343322754,
|
423 |
+
"5": 0.9110000729560852,
|
424 |
+
"6": 0.8740000128746033
|
425 |
+
},
|
426 |
+
"llm_top_1_test_accuracy": {
|
427 |
+
"1": 0.668,
|
428 |
+
"2": 0.573,
|
429 |
+
"3": 0.591,
|
430 |
+
"5": 0.568,
|
431 |
+
"6": 0.59
|
432 |
+
},
|
433 |
+
"llm_top_2_test_accuracy": {
|
434 |
+
"1": 0.745,
|
435 |
+
"2": 0.636,
|
436 |
+
"3": 0.606,
|
437 |
+
"5": 0.595,
|
438 |
+
"6": 0.621
|
439 |
+
},
|
440 |
+
"llm_top_5_test_accuracy": {
|
441 |
+
"1": 0.778,
|
442 |
+
"2": 0.643,
|
443 |
+
"3": 0.608,
|
444 |
+
"5": 0.66,
|
445 |
+
"6": 0.685
|
446 |
+
},
|
447 |
+
"sae_top_1_test_accuracy": {
|
448 |
+
"1": 0.839,
|
449 |
+
"2": 0.649,
|
450 |
+
"3": 0.739,
|
451 |
+
"5": 0.567,
|
452 |
+
"6": 0.781
|
453 |
+
},
|
454 |
+
"sae_top_2_test_accuracy": {
|
455 |
+
"1": 0.838,
|
456 |
+
"2": 0.727,
|
457 |
+
"3": 0.753,
|
458 |
+
"5": 0.866,
|
459 |
+
"6": 0.769
|
460 |
+
},
|
461 |
+
"sae_top_5_test_accuracy": {
|
462 |
+
"1": 0.857,
|
463 |
+
"2": 0.76,
|
464 |
+
"3": 0.763,
|
465 |
+
"5": 0.859,
|
466 |
+
"6": 0.772
|
467 |
+
}
|
468 |
+
},
|
469 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment_results": {
|
470 |
+
"sae_test_accuracy": {
|
471 |
+
"1.0": 0.9730000495910645,
|
472 |
+
"5.0": 0.9730000495910645
|
473 |
+
},
|
474 |
+
"llm_test_accuracy": {
|
475 |
+
"1.0": 0.9810000658035278,
|
476 |
+
"5.0": 0.9830000400543213
|
477 |
+
},
|
478 |
+
"llm_top_1_test_accuracy": {
|
479 |
+
"1.0": 0.672,
|
480 |
+
"5.0": 0.672
|
481 |
+
},
|
482 |
+
"llm_top_2_test_accuracy": {
|
483 |
+
"1.0": 0.724,
|
484 |
+
"5.0": 0.724
|
485 |
+
},
|
486 |
+
"llm_top_5_test_accuracy": {
|
487 |
+
"1.0": 0.766,
|
488 |
+
"5.0": 0.766
|
489 |
+
},
|
490 |
+
"sae_top_1_test_accuracy": {
|
491 |
+
"1.0": 0.714,
|
492 |
+
"5.0": 0.714
|
493 |
+
},
|
494 |
+
"sae_top_2_test_accuracy": {
|
495 |
+
"1.0": 0.857,
|
496 |
+
"5.0": 0.857
|
497 |
+
},
|
498 |
+
"sae_top_5_test_accuracy": {
|
499 |
+
"1.0": 0.882,
|
500 |
+
"5.0": 0.882
|
501 |
+
}
|
502 |
+
},
|
503 |
+
"codeparrot/github-code_results": {
|
504 |
+
"sae_test_accuracy": {
|
505 |
+
"C": 0.956000030040741,
|
506 |
+
"Python": 0.9820000529289246,
|
507 |
+
"HTML": 0.987000048160553,
|
508 |
+
"Java": 0.9570000171661377,
|
509 |
+
"PHP": 0.9530000686645508
|
510 |
+
},
|
511 |
+
"llm_test_accuracy": {
|
512 |
+
"C": 0.9510000348091125,
|
513 |
+
"Python": 0.987000048160553,
|
514 |
+
"HTML": 0.984000027179718,
|
515 |
+
"Java": 0.968000054359436,
|
516 |
+
"PHP": 0.9460000395774841
|
517 |
+
},
|
518 |
+
"llm_top_1_test_accuracy": {
|
519 |
+
"C": 0.655,
|
520 |
+
"Python": 0.661,
|
521 |
+
"HTML": 0.802,
|
522 |
+
"Java": 0.622,
|
523 |
+
"PHP": 0.594
|
524 |
+
},
|
525 |
+
"llm_top_2_test_accuracy": {
|
526 |
+
"C": 0.667,
|
527 |
+
"Python": 0.668,
|
528 |
+
"HTML": 0.798,
|
529 |
+
"Java": 0.674,
|
530 |
+
"PHP": 0.641
|
531 |
+
},
|
532 |
+
"llm_top_5_test_accuracy": {
|
533 |
+
"C": 0.78,
|
534 |
+
"Python": 0.73,
|
535 |
+
"HTML": 0.906,
|
536 |
+
"Java": 0.728,
|
537 |
+
"PHP": 0.684
|
538 |
+
},
|
539 |
+
"sae_top_1_test_accuracy": {
|
540 |
+
"C": 0.575,
|
541 |
+
"Python": 0.624,
|
542 |
+
"HTML": 0.924,
|
543 |
+
"Java": 0.66,
|
544 |
+
"PHP": 0.619
|
545 |
+
},
|
546 |
+
"sae_top_2_test_accuracy": {
|
547 |
+
"C": 0.663,
|
548 |
+
"Python": 0.606,
|
549 |
+
"HTML": 0.934,
|
550 |
+
"Java": 0.66,
|
551 |
+
"PHP": 0.634
|
552 |
+
},
|
553 |
+
"sae_top_5_test_accuracy": {
|
554 |
+
"C": 0.893,
|
555 |
+
"Python": 0.719,
|
556 |
+
"HTML": 0.938,
|
557 |
+
"Java": 0.73,
|
558 |
+
"PHP": 0.931
|
559 |
+
}
|
560 |
+
},
|
561 |
+
"fancyzhx/ag_news_results": {
|
562 |
+
"sae_test_accuracy": {
|
563 |
+
"0": 0.9440000653266907,
|
564 |
+
"1": 0.9830000400543213,
|
565 |
+
"2": 0.9320000410079956,
|
566 |
+
"3": 0.9450000524520874
|
567 |
+
},
|
568 |
+
"llm_test_accuracy": {
|
569 |
+
"0": 0.9360000491142273,
|
570 |
+
"1": 0.9890000224113464,
|
571 |
+
"2": 0.9280000329017639,
|
572 |
+
"3": 0.9500000476837158
|
573 |
+
},
|
574 |
+
"llm_top_1_test_accuracy": {
|
575 |
+
"0": 0.58,
|
576 |
+
"1": 0.676,
|
577 |
+
"2": 0.667,
|
578 |
+
"3": 0.622
|
579 |
+
},
|
580 |
+
"llm_top_2_test_accuracy": {
|
581 |
+
"0": 0.811,
|
582 |
+
"1": 0.798,
|
583 |
+
"2": 0.693,
|
584 |
+
"3": 0.76
|
585 |
+
},
|
586 |
+
"llm_top_5_test_accuracy": {
|
587 |
+
"0": 0.815,
|
588 |
+
"1": 0.884,
|
589 |
+
"2": 0.756,
|
590 |
+
"3": 0.844
|
591 |
+
},
|
592 |
+
"sae_top_1_test_accuracy": {
|
593 |
+
"0": 0.825,
|
594 |
+
"1": 0.943,
|
595 |
+
"2": 0.591,
|
596 |
+
"3": 0.721
|
597 |
+
},
|
598 |
+
"sae_top_2_test_accuracy": {
|
599 |
+
"0": 0.877,
|
600 |
+
"1": 0.943,
|
601 |
+
"2": 0.742,
|
602 |
+
"3": 0.863
|
603 |
+
},
|
604 |
+
"sae_top_5_test_accuracy": {
|
605 |
+
"0": 0.89,
|
606 |
+
"1": 0.97,
|
607 |
+
"2": 0.862,
|
608 |
+
"3": 0.891
|
609 |
+
}
|
610 |
+
},
|
611 |
+
"Helsinki-NLP/europarl_results": {
|
612 |
+
"sae_test_accuracy": {
|
613 |
+
"en": 0.9980000257492065,
|
614 |
+
"fr": 1.0,
|
615 |
+
"de": 1.0,
|
616 |
+
"es": 0.9980000257492065,
|
617 |
+
"nl": 1.0
|
618 |
+
},
|
619 |
+
"llm_test_accuracy": {
|
620 |
+
"en": 1.0,
|
621 |
+
"fr": 0.999000072479248,
|
622 |
+
"de": 0.999000072479248,
|
623 |
+
"es": 1.0,
|
624 |
+
"nl": 0.999000072479248
|
625 |
+
},
|
626 |
+
"llm_top_1_test_accuracy": {
|
627 |
+
"en": 0.748,
|
628 |
+
"fr": 0.615,
|
629 |
+
"de": 0.761,
|
630 |
+
"es": 0.904,
|
631 |
+
"nl": 0.657
|
632 |
+
},
|
633 |
+
"llm_top_2_test_accuracy": {
|
634 |
+
"en": 0.829,
|
635 |
+
"fr": 0.583,
|
636 |
+
"de": 0.829,
|
637 |
+
"es": 0.914,
|
638 |
+
"nl": 0.743
|
639 |
+
},
|
640 |
+
"llm_top_5_test_accuracy": {
|
641 |
+
"en": 0.892,
|
642 |
+
"fr": 0.914,
|
643 |
+
"de": 0.917,
|
644 |
+
"es": 0.984,
|
645 |
+
"nl": 0.856
|
646 |
+
},
|
647 |
+
"sae_top_1_test_accuracy": {
|
648 |
+
"en": 0.999,
|
649 |
+
"fr": 0.818,
|
650 |
+
"de": 0.884,
|
651 |
+
"es": 0.866,
|
652 |
+
"nl": 0.655
|
653 |
+
},
|
654 |
+
"sae_top_2_test_accuracy": {
|
655 |
+
"en": 0.999,
|
656 |
+
"fr": 0.991,
|
657 |
+
"de": 0.889,
|
658 |
+
"es": 0.994,
|
659 |
+
"nl": 0.883
|
660 |
+
},
|
661 |
+
"sae_top_5_test_accuracy": {
|
662 |
+
"en": 0.998,
|
663 |
+
"fr": 0.995,
|
664 |
+
"de": 0.969,
|
665 |
+
"es": 0.995,
|
666 |
+
"nl": 0.878
|
667 |
+
}
|
668 |
+
}
|
669 |
+
}
|
670 |
+
}
|
sparse_probing/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,670 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "sparse_probing",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"LabHC/bias_in_bios_class_set1",
|
7 |
+
"LabHC/bias_in_bios_class_set2",
|
8 |
+
"LabHC/bias_in_bios_class_set3",
|
9 |
+
"canrager/amazon_reviews_mcauley_1and5",
|
10 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment",
|
11 |
+
"codeparrot/github-code",
|
12 |
+
"fancyzhx/ag_news",
|
13 |
+
"Helsinki-NLP/europarl"
|
14 |
+
],
|
15 |
+
"probe_train_set_size": 4000,
|
16 |
+
"probe_test_set_size": 1000,
|
17 |
+
"context_length": 128,
|
18 |
+
"sae_batch_size": 125,
|
19 |
+
"llm_batch_size": 32,
|
20 |
+
"llm_dtype": "bfloat16",
|
21 |
+
"model_name": "gemma-2-2b",
|
22 |
+
"k_values": [
|
23 |
+
1,
|
24 |
+
2,
|
25 |
+
5
|
26 |
+
],
|
27 |
+
"lower_vram_usage": false
|
28 |
+
},
|
29 |
+
"eval_id": "cc7d3f8e-6433-4738-b99d-424001238fe9",
|
30 |
+
"datetime_epoch_millis": 1737824318467,
|
31 |
+
"eval_result_metrics": {
|
32 |
+
"llm": {
|
33 |
+
"llm_test_accuracy": 0.9590937938541175,
|
34 |
+
"llm_top_1_test_accuracy": 0.66480625,
|
35 |
+
"llm_top_2_test_accuracy": 0.7173375000000001,
|
36 |
+
"llm_top_5_test_accuracy": 0.78036875,
|
37 |
+
"llm_top_10_test_accuracy": null,
|
38 |
+
"llm_top_20_test_accuracy": null,
|
39 |
+
"llm_top_50_test_accuracy": null,
|
40 |
+
"llm_top_100_test_accuracy": null
|
41 |
+
},
|
42 |
+
"sae": {
|
43 |
+
"sae_test_accuracy": 0.957118796557188,
|
44 |
+
"sae_top_1_test_accuracy": 0.7846500000000001,
|
45 |
+
"sae_top_2_test_accuracy": 0.8207937500000001,
|
46 |
+
"sae_top_5_test_accuracy": 0.85040625,
|
47 |
+
"sae_top_10_test_accuracy": null,
|
48 |
+
"sae_top_20_test_accuracy": null,
|
49 |
+
"sae_top_50_test_accuracy": null,
|
50 |
+
"sae_top_100_test_accuracy": null
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"eval_result_details": [
|
54 |
+
{
|
55 |
+
"dataset_name": "LabHC/bias_in_bios_class_set1_results",
|
56 |
+
"llm_test_accuracy": 0.9694000363349915,
|
57 |
+
"llm_top_1_test_accuracy": 0.6436000000000001,
|
58 |
+
"llm_top_2_test_accuracy": 0.6874,
|
59 |
+
"llm_top_5_test_accuracy": 0.7908,
|
60 |
+
"llm_top_10_test_accuracy": null,
|
61 |
+
"llm_top_20_test_accuracy": null,
|
62 |
+
"llm_top_50_test_accuracy": null,
|
63 |
+
"llm_top_100_test_accuracy": null,
|
64 |
+
"sae_test_accuracy": 0.965600049495697,
|
65 |
+
"sae_top_1_test_accuracy": 0.7788,
|
66 |
+
"sae_top_2_test_accuracy": 0.7933999999999999,
|
67 |
+
"sae_top_5_test_accuracy": 0.857,
|
68 |
+
"sae_top_10_test_accuracy": null,
|
69 |
+
"sae_top_20_test_accuracy": null,
|
70 |
+
"sae_top_50_test_accuracy": null,
|
71 |
+
"sae_top_100_test_accuracy": null
|
72 |
+
},
|
73 |
+
{
|
74 |
+
"dataset_name": "LabHC/bias_in_bios_class_set2_results",
|
75 |
+
"llm_test_accuracy": 0.9544000387191772,
|
76 |
+
"llm_top_1_test_accuracy": 0.6764,
|
77 |
+
"llm_top_2_test_accuracy": 0.7150000000000001,
|
78 |
+
"llm_top_5_test_accuracy": 0.7592000000000001,
|
79 |
+
"llm_top_10_test_accuracy": null,
|
80 |
+
"llm_top_20_test_accuracy": null,
|
81 |
+
"llm_top_50_test_accuracy": null,
|
82 |
+
"llm_top_100_test_accuracy": null,
|
83 |
+
"sae_test_accuracy": 0.9498000502586365,
|
84 |
+
"sae_top_1_test_accuracy": 0.7714,
|
85 |
+
"sae_top_2_test_accuracy": 0.7968,
|
86 |
+
"sae_top_5_test_accuracy": 0.825,
|
87 |
+
"sae_top_10_test_accuracy": null,
|
88 |
+
"sae_top_20_test_accuracy": null,
|
89 |
+
"sae_top_50_test_accuracy": null,
|
90 |
+
"sae_top_100_test_accuracy": null
|
91 |
+
},
|
92 |
+
{
|
93 |
+
"dataset_name": "LabHC/bias_in_bios_class_set3_results",
|
94 |
+
"llm_test_accuracy": 0.9332000494003296,
|
95 |
+
"llm_top_1_test_accuracy": 0.6884,
|
96 |
+
"llm_top_2_test_accuracy": 0.737,
|
97 |
+
"llm_top_5_test_accuracy": 0.7492,
|
98 |
+
"llm_top_10_test_accuracy": null,
|
99 |
+
"llm_top_20_test_accuracy": null,
|
100 |
+
"llm_top_50_test_accuracy": null,
|
101 |
+
"llm_top_100_test_accuracy": null,
|
102 |
+
"sae_test_accuracy": 0.9310000658035278,
|
103 |
+
"sae_top_1_test_accuracy": 0.7462,
|
104 |
+
"sae_top_2_test_accuracy": 0.7996000000000001,
|
105 |
+
"sae_top_5_test_accuracy": 0.8216000000000001,
|
106 |
+
"sae_top_10_test_accuracy": null,
|
107 |
+
"sae_top_20_test_accuracy": null,
|
108 |
+
"sae_top_50_test_accuracy": null,
|
109 |
+
"sae_top_100_test_accuracy": null
|
110 |
+
},
|
111 |
+
{
|
112 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_results",
|
113 |
+
"llm_test_accuracy": 0.9164000511169433,
|
114 |
+
"llm_top_1_test_accuracy": 0.598,
|
115 |
+
"llm_top_2_test_accuracy": 0.6406,
|
116 |
+
"llm_top_5_test_accuracy": 0.6748000000000001,
|
117 |
+
"llm_top_10_test_accuracy": null,
|
118 |
+
"llm_top_20_test_accuracy": null,
|
119 |
+
"llm_top_50_test_accuracy": null,
|
120 |
+
"llm_top_100_test_accuracy": null,
|
121 |
+
"sae_test_accuracy": 0.9112000465393066,
|
122 |
+
"sae_top_1_test_accuracy": 0.6738,
|
123 |
+
"sae_top_2_test_accuracy": 0.7656,
|
124 |
+
"sae_top_5_test_accuracy": 0.7866,
|
125 |
+
"sae_top_10_test_accuracy": null,
|
126 |
+
"sae_top_20_test_accuracy": null,
|
127 |
+
"sae_top_50_test_accuracy": null,
|
128 |
+
"sae_top_100_test_accuracy": null
|
129 |
+
},
|
130 |
+
{
|
131 |
+
"dataset_name": "canrager/amazon_reviews_mcauley_1and5_sentiment_results",
|
132 |
+
"llm_test_accuracy": 0.9820000529289246,
|
133 |
+
"llm_top_1_test_accuracy": 0.672,
|
134 |
+
"llm_top_2_test_accuracy": 0.724,
|
135 |
+
"llm_top_5_test_accuracy": 0.766,
|
136 |
+
"llm_top_10_test_accuracy": null,
|
137 |
+
"llm_top_20_test_accuracy": null,
|
138 |
+
"llm_top_50_test_accuracy": null,
|
139 |
+
"llm_top_100_test_accuracy": null,
|
140 |
+
"sae_test_accuracy": 0.9780000448226929,
|
141 |
+
"sae_top_1_test_accuracy": 0.906,
|
142 |
+
"sae_top_2_test_accuracy": 0.905,
|
143 |
+
"sae_top_5_test_accuracy": 0.906,
|
144 |
+
"sae_top_10_test_accuracy": null,
|
145 |
+
"sae_top_20_test_accuracy": null,
|
146 |
+
"sae_top_50_test_accuracy": null,
|
147 |
+
"sae_top_100_test_accuracy": null
|
148 |
+
},
|
149 |
+
{
|
150 |
+
"dataset_name": "codeparrot/github-code_results",
|
151 |
+
"llm_test_accuracy": 0.9672000408172607,
|
152 |
+
"llm_top_1_test_accuracy": 0.6668000000000001,
|
153 |
+
"llm_top_2_test_accuracy": 0.6896,
|
154 |
+
"llm_top_5_test_accuracy": 0.7656000000000001,
|
155 |
+
"llm_top_10_test_accuracy": null,
|
156 |
+
"llm_top_20_test_accuracy": null,
|
157 |
+
"llm_top_50_test_accuracy": null,
|
158 |
+
"llm_top_100_test_accuracy": null,
|
159 |
+
"sae_test_accuracy": 0.9688000440597534,
|
160 |
+
"sae_top_1_test_accuracy": 0.676,
|
161 |
+
"sae_top_2_test_accuracy": 0.7224,
|
162 |
+
"sae_top_5_test_accuracy": 0.7539999999999999,
|
163 |
+
"sae_top_10_test_accuracy": null,
|
164 |
+
"sae_top_20_test_accuracy": null,
|
165 |
+
"sae_top_50_test_accuracy": null,
|
166 |
+
"sae_top_100_test_accuracy": null
|
167 |
+
},
|
168 |
+
{
|
169 |
+
"dataset_name": "fancyzhx/ag_news_results",
|
170 |
+
"llm_test_accuracy": 0.9507500380277634,
|
171 |
+
"llm_top_1_test_accuracy": 0.63625,
|
172 |
+
"llm_top_2_test_accuracy": 0.7655000000000001,
|
173 |
+
"llm_top_5_test_accuracy": 0.82475,
|
174 |
+
"llm_top_10_test_accuracy": null,
|
175 |
+
"llm_top_20_test_accuracy": null,
|
176 |
+
"llm_top_50_test_accuracy": null,
|
177 |
+
"llm_top_100_test_accuracy": null,
|
178 |
+
"sae_test_accuracy": 0.9527500569820404,
|
179 |
+
"sae_top_1_test_accuracy": 0.841,
|
180 |
+
"sae_top_2_test_accuracy": 0.88475,
|
181 |
+
"sae_top_5_test_accuracy": 0.89925,
|
182 |
+
"sae_top_10_test_accuracy": null,
|
183 |
+
"sae_top_20_test_accuracy": null,
|
184 |
+
"sae_top_50_test_accuracy": null,
|
185 |
+
"sae_top_100_test_accuracy": null
|
186 |
+
},
|
187 |
+
{
|
188 |
+
"dataset_name": "Helsinki-NLP/europarl_results",
|
189 |
+
"llm_test_accuracy": 0.9994000434875489,
|
190 |
+
"llm_top_1_test_accuracy": 0.737,
|
191 |
+
"llm_top_2_test_accuracy": 0.7796,
|
192 |
+
"llm_top_5_test_accuracy": 0.9126,
|
193 |
+
"llm_top_10_test_accuracy": null,
|
194 |
+
"llm_top_20_test_accuracy": null,
|
195 |
+
"llm_top_50_test_accuracy": null,
|
196 |
+
"llm_top_100_test_accuracy": null,
|
197 |
+
"sae_test_accuracy": 0.9998000144958497,
|
198 |
+
"sae_top_1_test_accuracy": 0.884,
|
199 |
+
"sae_top_2_test_accuracy": 0.8987999999999999,
|
200 |
+
"sae_top_5_test_accuracy": 0.9538,
|
201 |
+
"sae_top_10_test_accuracy": null,
|
202 |
+
"sae_top_20_test_accuracy": null,
|
203 |
+
"sae_top_50_test_accuracy": null,
|
204 |
+
"sae_top_100_test_accuracy": null
|
205 |
+
}
|
206 |
+
],
|
207 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
208 |
+
"sae_lens_id": "custom_sae",
|
209 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5",
|
210 |
+
"sae_lens_version": "5.3.2",
|
211 |
+
"sae_cfg_dict": {
|
212 |
+
"model_name": "gemma-2-2b",
|
213 |
+
"d_in": 2304,
|
214 |
+
"d_sae": 16384,
|
215 |
+
"hook_layer": 12,
|
216 |
+
"hook_name": "blocks.12.hook_resid_post",
|
217 |
+
"context_size": null,
|
218 |
+
"hook_head_index": null,
|
219 |
+
"architecture": "matryoshka_batch_topk",
|
220 |
+
"apply_b_dec_to_input": null,
|
221 |
+
"finetuning_scaling_factor": null,
|
222 |
+
"activation_fn_str": "",
|
223 |
+
"prepend_bos": true,
|
224 |
+
"normalize_activations": "none",
|
225 |
+
"dtype": "bfloat16",
|
226 |
+
"device": "",
|
227 |
+
"dataset_path": "",
|
228 |
+
"dataset_trust_remote_code": true,
|
229 |
+
"seqpos_slice": [
|
230 |
+
null
|
231 |
+
],
|
232 |
+
"training_tokens": -100000,
|
233 |
+
"sae_lens_training_version": null,
|
234 |
+
"neuronpedia_id": null
|
235 |
+
},
|
236 |
+
"eval_result_unstructured": {
|
237 |
+
"LabHC/bias_in_bios_class_set1_results": {
|
238 |
+
"sae_test_accuracy": {
|
239 |
+
"0": 0.9480000734329224,
|
240 |
+
"1": 0.956000030040741,
|
241 |
+
"2": 0.9590000510215759,
|
242 |
+
"6": 0.9880000352859497,
|
243 |
+
"9": 0.9770000576972961
|
244 |
+
},
|
245 |
+
"llm_test_accuracy": {
|
246 |
+
"0": 0.9510000348091125,
|
247 |
+
"1": 0.9670000672340393,
|
248 |
+
"2": 0.9520000219345093,
|
249 |
+
"6": 0.9930000305175781,
|
250 |
+
"9": 0.984000027179718
|
251 |
+
},
|
252 |
+
"llm_top_1_test_accuracy": {
|
253 |
+
"0": 0.568,
|
254 |
+
"1": 0.629,
|
255 |
+
"2": 0.679,
|
256 |
+
"6": 0.791,
|
257 |
+
"9": 0.551
|
258 |
+
},
|
259 |
+
"llm_top_2_test_accuracy": {
|
260 |
+
"0": 0.585,
|
261 |
+
"1": 0.666,
|
262 |
+
"2": 0.673,
|
263 |
+
"6": 0.801,
|
264 |
+
"9": 0.712
|
265 |
+
},
|
266 |
+
"llm_top_5_test_accuracy": {
|
267 |
+
"0": 0.72,
|
268 |
+
"1": 0.707,
|
269 |
+
"2": 0.764,
|
270 |
+
"6": 0.899,
|
271 |
+
"9": 0.864
|
272 |
+
},
|
273 |
+
"sae_top_1_test_accuracy": {
|
274 |
+
"0": 0.642,
|
275 |
+
"1": 0.679,
|
276 |
+
"2": 0.681,
|
277 |
+
"6": 0.956,
|
278 |
+
"9": 0.936
|
279 |
+
},
|
280 |
+
"sae_top_2_test_accuracy": {
|
281 |
+
"0": 0.647,
|
282 |
+
"1": 0.692,
|
283 |
+
"2": 0.717,
|
284 |
+
"6": 0.97,
|
285 |
+
"9": 0.941
|
286 |
+
},
|
287 |
+
"sae_top_5_test_accuracy": {
|
288 |
+
"0": 0.747,
|
289 |
+
"1": 0.738,
|
290 |
+
"2": 0.868,
|
291 |
+
"6": 0.982,
|
292 |
+
"9": 0.95
|
293 |
+
}
|
294 |
+
},
|
295 |
+
"LabHC/bias_in_bios_class_set2_results": {
|
296 |
+
"sae_test_accuracy": {
|
297 |
+
"11": 0.9590000510215759,
|
298 |
+
"13": 0.9470000267028809,
|
299 |
+
"14": 0.9580000638961792,
|
300 |
+
"18": 0.9270000457763672,
|
301 |
+
"19": 0.9580000638961792
|
302 |
+
},
|
303 |
+
"llm_test_accuracy": {
|
304 |
+
"11": 0.9650000333786011,
|
305 |
+
"13": 0.9520000219345093,
|
306 |
+
"14": 0.9550000429153442,
|
307 |
+
"18": 0.9360000491142273,
|
308 |
+
"19": 0.9640000462532043
|
309 |
+
},
|
310 |
+
"llm_top_1_test_accuracy": {
|
311 |
+
"11": 0.558,
|
312 |
+
"13": 0.673,
|
313 |
+
"14": 0.656,
|
314 |
+
"18": 0.702,
|
315 |
+
"19": 0.793
|
316 |
+
},
|
317 |
+
"llm_top_2_test_accuracy": {
|
318 |
+
"11": 0.686,
|
319 |
+
"13": 0.713,
|
320 |
+
"14": 0.687,
|
321 |
+
"18": 0.724,
|
322 |
+
"19": 0.765
|
323 |
+
},
|
324 |
+
"llm_top_5_test_accuracy": {
|
325 |
+
"11": 0.782,
|
326 |
+
"13": 0.742,
|
327 |
+
"14": 0.716,
|
328 |
+
"18": 0.725,
|
329 |
+
"19": 0.831
|
330 |
+
},
|
331 |
+
"sae_top_1_test_accuracy": {
|
332 |
+
"11": 0.736,
|
333 |
+
"13": 0.686,
|
334 |
+
"14": 0.828,
|
335 |
+
"18": 0.768,
|
336 |
+
"19": 0.839
|
337 |
+
},
|
338 |
+
"sae_top_2_test_accuracy": {
|
339 |
+
"11": 0.759,
|
340 |
+
"13": 0.763,
|
341 |
+
"14": 0.838,
|
342 |
+
"18": 0.755,
|
343 |
+
"19": 0.869
|
344 |
+
},
|
345 |
+
"sae_top_5_test_accuracy": {
|
346 |
+
"11": 0.845,
|
347 |
+
"13": 0.813,
|
348 |
+
"14": 0.834,
|
349 |
+
"18": 0.757,
|
350 |
+
"19": 0.876
|
351 |
+
}
|
352 |
+
},
|
353 |
+
"LabHC/bias_in_bios_class_set3_results": {
|
354 |
+
"sae_test_accuracy": {
|
355 |
+
"20": 0.9620000720024109,
|
356 |
+
"21": 0.9300000667572021,
|
357 |
+
"22": 0.921000063419342,
|
358 |
+
"25": 0.9530000686645508,
|
359 |
+
"26": 0.8890000581741333
|
360 |
+
},
|
361 |
+
"llm_test_accuracy": {
|
362 |
+
"20": 0.9610000252723694,
|
363 |
+
"21": 0.9270000457763672,
|
364 |
+
"22": 0.921000063419342,
|
365 |
+
"25": 0.968000054359436,
|
366 |
+
"26": 0.8890000581741333
|
367 |
+
},
|
368 |
+
"llm_top_1_test_accuracy": {
|
369 |
+
"20": 0.713,
|
370 |
+
"21": 0.766,
|
371 |
+
"22": 0.639,
|
372 |
+
"25": 0.69,
|
373 |
+
"26": 0.634
|
374 |
+
},
|
375 |
+
"llm_top_2_test_accuracy": {
|
376 |
+
"20": 0.804,
|
377 |
+
"21": 0.777,
|
378 |
+
"22": 0.68,
|
379 |
+
"25": 0.759,
|
380 |
+
"26": 0.665
|
381 |
+
},
|
382 |
+
"llm_top_5_test_accuracy": {
|
383 |
+
"20": 0.813,
|
384 |
+
"21": 0.777,
|
385 |
+
"22": 0.686,
|
386 |
+
"25": 0.798,
|
387 |
+
"26": 0.672
|
388 |
+
},
|
389 |
+
"sae_top_1_test_accuracy": {
|
390 |
+
"20": 0.717,
|
391 |
+
"21": 0.78,
|
392 |
+
"22": 0.729,
|
393 |
+
"25": 0.856,
|
394 |
+
"26": 0.649
|
395 |
+
},
|
396 |
+
"sae_top_2_test_accuracy": {
|
397 |
+
"20": 0.895,
|
398 |
+
"21": 0.779,
|
399 |
+
"22": 0.729,
|
400 |
+
"25": 0.862,
|
401 |
+
"26": 0.733
|
402 |
+
},
|
403 |
+
"sae_top_5_test_accuracy": {
|
404 |
+
"20": 0.898,
|
405 |
+
"21": 0.838,
|
406 |
+
"22": 0.745,
|
407 |
+
"25": 0.889,
|
408 |
+
"26": 0.738
|
409 |
+
}
|
410 |
+
},
|
411 |
+
"canrager/amazon_reviews_mcauley_1and5_results": {
|
412 |
+
"sae_test_accuracy": {
|
413 |
+
"1": 0.9500000476837158,
|
414 |
+
"2": 0.9340000152587891,
|
415 |
+
"3": 0.9130000472068787,
|
416 |
+
"5": 0.9170000553131104,
|
417 |
+
"6": 0.8420000672340393
|
418 |
+
},
|
419 |
+
"llm_test_accuracy": {
|
420 |
+
"1": 0.9480000734329224,
|
421 |
+
"2": 0.9350000619888306,
|
422 |
+
"3": 0.9140000343322754,
|
423 |
+
"5": 0.9110000729560852,
|
424 |
+
"6": 0.8740000128746033
|
425 |
+
},
|
426 |
+
"llm_top_1_test_accuracy": {
|
427 |
+
"1": 0.668,
|
428 |
+
"2": 0.573,
|
429 |
+
"3": 0.591,
|
430 |
+
"5": 0.568,
|
431 |
+
"6": 0.59
|
432 |
+
},
|
433 |
+
"llm_top_2_test_accuracy": {
|
434 |
+
"1": 0.745,
|
435 |
+
"2": 0.636,
|
436 |
+
"3": 0.606,
|
437 |
+
"5": 0.595,
|
438 |
+
"6": 0.621
|
439 |
+
},
|
440 |
+
"llm_top_5_test_accuracy": {
|
441 |
+
"1": 0.778,
|
442 |
+
"2": 0.643,
|
443 |
+
"3": 0.608,
|
444 |
+
"5": 0.66,
|
445 |
+
"6": 0.685
|
446 |
+
},
|
447 |
+
"sae_top_1_test_accuracy": {
|
448 |
+
"1": 0.746,
|
449 |
+
"2": 0.643,
|
450 |
+
"3": 0.726,
|
451 |
+
"5": 0.558,
|
452 |
+
"6": 0.696
|
453 |
+
},
|
454 |
+
"sae_top_2_test_accuracy": {
|
455 |
+
"1": 0.833,
|
456 |
+
"2": 0.667,
|
457 |
+
"3": 0.743,
|
458 |
+
"5": 0.843,
|
459 |
+
"6": 0.742
|
460 |
+
},
|
461 |
+
"sae_top_5_test_accuracy": {
|
462 |
+
"1": 0.893,
|
463 |
+
"2": 0.705,
|
464 |
+
"3": 0.739,
|
465 |
+
"5": 0.838,
|
466 |
+
"6": 0.758
|
467 |
+
}
|
468 |
+
},
|
469 |
+
"canrager/amazon_reviews_mcauley_1and5_sentiment_results": {
|
470 |
+
"sae_test_accuracy": {
|
471 |
+
"1.0": 0.9780000448226929,
|
472 |
+
"5.0": 0.9780000448226929
|
473 |
+
},
|
474 |
+
"llm_test_accuracy": {
|
475 |
+
"1.0": 0.9810000658035278,
|
476 |
+
"5.0": 0.9830000400543213
|
477 |
+
},
|
478 |
+
"llm_top_1_test_accuracy": {
|
479 |
+
"1.0": 0.672,
|
480 |
+
"5.0": 0.672
|
481 |
+
},
|
482 |
+
"llm_top_2_test_accuracy": {
|
483 |
+
"1.0": 0.724,
|
484 |
+
"5.0": 0.724
|
485 |
+
},
|
486 |
+
"llm_top_5_test_accuracy": {
|
487 |
+
"1.0": 0.766,
|
488 |
+
"5.0": 0.766
|
489 |
+
},
|
490 |
+
"sae_top_1_test_accuracy": {
|
491 |
+
"1.0": 0.906,
|
492 |
+
"5.0": 0.906
|
493 |
+
},
|
494 |
+
"sae_top_2_test_accuracy": {
|
495 |
+
"1.0": 0.905,
|
496 |
+
"5.0": 0.905
|
497 |
+
},
|
498 |
+
"sae_top_5_test_accuracy": {
|
499 |
+
"1.0": 0.906,
|
500 |
+
"5.0": 0.906
|
501 |
+
}
|
502 |
+
},
|
503 |
+
"codeparrot/github-code_results": {
|
504 |
+
"sae_test_accuracy": {
|
505 |
+
"C": 0.9530000686645508,
|
506 |
+
"Python": 0.9820000529289246,
|
507 |
+
"HTML": 0.987000048160553,
|
508 |
+
"Java": 0.9650000333786011,
|
509 |
+
"PHP": 0.9570000171661377
|
510 |
+
},
|
511 |
+
"llm_test_accuracy": {
|
512 |
+
"C": 0.9510000348091125,
|
513 |
+
"Python": 0.987000048160553,
|
514 |
+
"HTML": 0.984000027179718,
|
515 |
+
"Java": 0.968000054359436,
|
516 |
+
"PHP": 0.9460000395774841
|
517 |
+
},
|
518 |
+
"llm_top_1_test_accuracy": {
|
519 |
+
"C": 0.655,
|
520 |
+
"Python": 0.661,
|
521 |
+
"HTML": 0.802,
|
522 |
+
"Java": 0.622,
|
523 |
+
"PHP": 0.594
|
524 |
+
},
|
525 |
+
"llm_top_2_test_accuracy": {
|
526 |
+
"C": 0.667,
|
527 |
+
"Python": 0.668,
|
528 |
+
"HTML": 0.798,
|
529 |
+
"Java": 0.674,
|
530 |
+
"PHP": 0.641
|
531 |
+
},
|
532 |
+
"llm_top_5_test_accuracy": {
|
533 |
+
"C": 0.78,
|
534 |
+
"Python": 0.73,
|
535 |
+
"HTML": 0.906,
|
536 |
+
"Java": 0.728,
|
537 |
+
"PHP": 0.684
|
538 |
+
},
|
539 |
+
"sae_top_1_test_accuracy": {
|
540 |
+
"C": 0.644,
|
541 |
+
"Python": 0.633,
|
542 |
+
"HTML": 0.85,
|
543 |
+
"Java": 0.644,
|
544 |
+
"PHP": 0.609
|
545 |
+
},
|
546 |
+
"sae_top_2_test_accuracy": {
|
547 |
+
"C": 0.747,
|
548 |
+
"Python": 0.655,
|
549 |
+
"HTML": 0.898,
|
550 |
+
"Java": 0.695,
|
551 |
+
"PHP": 0.617
|
552 |
+
},
|
553 |
+
"sae_top_5_test_accuracy": {
|
554 |
+
"C": 0.764,
|
555 |
+
"Python": 0.675,
|
556 |
+
"HTML": 0.909,
|
557 |
+
"Java": 0.723,
|
558 |
+
"PHP": 0.699
|
559 |
+
}
|
560 |
+
},
|
561 |
+
"fancyzhx/ag_news_results": {
|
562 |
+
"sae_test_accuracy": {
|
563 |
+
"0": 0.9440000653266907,
|
564 |
+
"1": 0.987000048160553,
|
565 |
+
"2": 0.9300000667572021,
|
566 |
+
"3": 0.9500000476837158
|
567 |
+
},
|
568 |
+
"llm_test_accuracy": {
|
569 |
+
"0": 0.9360000491142273,
|
570 |
+
"1": 0.9890000224113464,
|
571 |
+
"2": 0.9280000329017639,
|
572 |
+
"3": 0.9500000476837158
|
573 |
+
},
|
574 |
+
"llm_top_1_test_accuracy": {
|
575 |
+
"0": 0.58,
|
576 |
+
"1": 0.676,
|
577 |
+
"2": 0.667,
|
578 |
+
"3": 0.622
|
579 |
+
},
|
580 |
+
"llm_top_2_test_accuracy": {
|
581 |
+
"0": 0.811,
|
582 |
+
"1": 0.798,
|
583 |
+
"2": 0.693,
|
584 |
+
"3": 0.76
|
585 |
+
},
|
586 |
+
"llm_top_5_test_accuracy": {
|
587 |
+
"0": 0.815,
|
588 |
+
"1": 0.884,
|
589 |
+
"2": 0.756,
|
590 |
+
"3": 0.844
|
591 |
+
},
|
592 |
+
"sae_top_1_test_accuracy": {
|
593 |
+
"0": 0.837,
|
594 |
+
"1": 0.968,
|
595 |
+
"2": 0.836,
|
596 |
+
"3": 0.723
|
597 |
+
},
|
598 |
+
"sae_top_2_test_accuracy": {
|
599 |
+
"0": 0.872,
|
600 |
+
"1": 0.966,
|
601 |
+
"2": 0.833,
|
602 |
+
"3": 0.868
|
603 |
+
},
|
604 |
+
"sae_top_5_test_accuracy": {
|
605 |
+
"0": 0.885,
|
606 |
+
"1": 0.968,
|
607 |
+
"2": 0.865,
|
608 |
+
"3": 0.879
|
609 |
+
}
|
610 |
+
},
|
611 |
+
"Helsinki-NLP/europarl_results": {
|
612 |
+
"sae_test_accuracy": {
|
613 |
+
"en": 1.0,
|
614 |
+
"fr": 1.0,
|
615 |
+
"de": 1.0,
|
616 |
+
"es": 0.999000072479248,
|
617 |
+
"nl": 1.0
|
618 |
+
},
|
619 |
+
"llm_test_accuracy": {
|
620 |
+
"en": 1.0,
|
621 |
+
"fr": 0.999000072479248,
|
622 |
+
"de": 0.999000072479248,
|
623 |
+
"es": 1.0,
|
624 |
+
"nl": 0.999000072479248
|
625 |
+
},
|
626 |
+
"llm_top_1_test_accuracy": {
|
627 |
+
"en": 0.748,
|
628 |
+
"fr": 0.615,
|
629 |
+
"de": 0.761,
|
630 |
+
"es": 0.904,
|
631 |
+
"nl": 0.657
|
632 |
+
},
|
633 |
+
"llm_top_2_test_accuracy": {
|
634 |
+
"en": 0.829,
|
635 |
+
"fr": 0.583,
|
636 |
+
"de": 0.829,
|
637 |
+
"es": 0.914,
|
638 |
+
"nl": 0.743
|
639 |
+
},
|
640 |
+
"llm_top_5_test_accuracy": {
|
641 |
+
"en": 0.892,
|
642 |
+
"fr": 0.914,
|
643 |
+
"de": 0.917,
|
644 |
+
"es": 0.984,
|
645 |
+
"nl": 0.856
|
646 |
+
},
|
647 |
+
"sae_top_1_test_accuracy": {
|
648 |
+
"en": 0.997,
|
649 |
+
"fr": 0.805,
|
650 |
+
"de": 0.881,
|
651 |
+
"es": 0.891,
|
652 |
+
"nl": 0.846
|
653 |
+
},
|
654 |
+
"sae_top_2_test_accuracy": {
|
655 |
+
"en": 0.997,
|
656 |
+
"fr": 0.826,
|
657 |
+
"de": 0.871,
|
658 |
+
"es": 0.945,
|
659 |
+
"nl": 0.855
|
660 |
+
},
|
661 |
+
"sae_top_5_test_accuracy": {
|
662 |
+
"en": 0.999,
|
663 |
+
"fr": 0.991,
|
664 |
+
"de": 0.948,
|
665 |
+
"es": 0.966,
|
666 |
+
"nl": 0.865
|
667 |
+
}
|
668 |
+
}
|
669 |
+
}
|
670 |
+
}
|
unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "unlearning",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"wmdp-bio",
|
7 |
+
"high_school_us_history",
|
8 |
+
"college_computer_science",
|
9 |
+
"high_school_geography",
|
10 |
+
"human_aging"
|
11 |
+
],
|
12 |
+
"intervention_method": "clamp_feature_activation",
|
13 |
+
"retain_thresholds": [
|
14 |
+
0.001,
|
15 |
+
0.01
|
16 |
+
],
|
17 |
+
"n_features_list": [
|
18 |
+
10,
|
19 |
+
20
|
20 |
+
],
|
21 |
+
"multipliers": [
|
22 |
+
25,
|
23 |
+
50,
|
24 |
+
100,
|
25 |
+
200
|
26 |
+
],
|
27 |
+
"dataset_size": 1024,
|
28 |
+
"seq_len": 1024,
|
29 |
+
"n_batch_loss_added": 50,
|
30 |
+
"target_metric": "correct",
|
31 |
+
"save_metrics": true,
|
32 |
+
"model_name": "gemma-2-2b-it",
|
33 |
+
"llm_batch_size": 4,
|
34 |
+
"llm_dtype": "bfloat16"
|
35 |
+
},
|
36 |
+
"eval_id": "44871cba-565b-4e6a-a1c4-856b6fcff9db",
|
37 |
+
"datetime_epoch_millis": 1737828354899,
|
38 |
+
"eval_result_metrics": {
|
39 |
+
"unlearning": {
|
40 |
+
"unlearning_score": 0.054409027099609375
|
41 |
+
}
|
42 |
+
},
|
43 |
+
"eval_result_details": [],
|
44 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
45 |
+
"sae_lens_id": "custom_sae",
|
46 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_0",
|
47 |
+
"sae_lens_version": "5.3.2",
|
48 |
+
"sae_cfg_dict": {
|
49 |
+
"model_name": "gemma-2-2b",
|
50 |
+
"d_in": 2304,
|
51 |
+
"d_sae": 16384,
|
52 |
+
"hook_layer": 12,
|
53 |
+
"hook_name": "blocks.12.hook_resid_post",
|
54 |
+
"context_size": null,
|
55 |
+
"hook_head_index": null,
|
56 |
+
"architecture": "matryoshka_batch_topk",
|
57 |
+
"apply_b_dec_to_input": null,
|
58 |
+
"finetuning_scaling_factor": null,
|
59 |
+
"activation_fn_str": "",
|
60 |
+
"prepend_bos": true,
|
61 |
+
"normalize_activations": "none",
|
62 |
+
"dtype": "bfloat16",
|
63 |
+
"device": "",
|
64 |
+
"dataset_path": "",
|
65 |
+
"dataset_trust_remote_code": true,
|
66 |
+
"seqpos_slice": [
|
67 |
+
null
|
68 |
+
],
|
69 |
+
"training_tokens": -100000,
|
70 |
+
"sae_lens_training_version": null,
|
71 |
+
"neuronpedia_id": null
|
72 |
+
},
|
73 |
+
"eval_result_unstructured": null
|
74 |
+
}
|
unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "unlearning",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"wmdp-bio",
|
7 |
+
"high_school_us_history",
|
8 |
+
"college_computer_science",
|
9 |
+
"high_school_geography",
|
10 |
+
"human_aging"
|
11 |
+
],
|
12 |
+
"intervention_method": "clamp_feature_activation",
|
13 |
+
"retain_thresholds": [
|
14 |
+
0.001,
|
15 |
+
0.01
|
16 |
+
],
|
17 |
+
"n_features_list": [
|
18 |
+
10,
|
19 |
+
20
|
20 |
+
],
|
21 |
+
"multipliers": [
|
22 |
+
25,
|
23 |
+
50,
|
24 |
+
100,
|
25 |
+
200
|
26 |
+
],
|
27 |
+
"dataset_size": 1024,
|
28 |
+
"seq_len": 1024,
|
29 |
+
"n_batch_loss_added": 50,
|
30 |
+
"target_metric": "correct",
|
31 |
+
"save_metrics": true,
|
32 |
+
"model_name": "gemma-2-2b-it",
|
33 |
+
"llm_batch_size": 4,
|
34 |
+
"llm_dtype": "bfloat16"
|
35 |
+
},
|
36 |
+
"eval_id": "3d8f8cec-cb25-492c-8b76-185b501a556c",
|
37 |
+
"datetime_epoch_millis": 1737829042723,
|
38 |
+
"eval_result_metrics": {
|
39 |
+
"unlearning": {
|
40 |
+
"unlearning_score": 0.04690432548522949
|
41 |
+
}
|
42 |
+
},
|
43 |
+
"eval_result_details": [],
|
44 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
45 |
+
"sae_lens_id": "custom_sae",
|
46 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_1",
|
47 |
+
"sae_lens_version": "5.3.2",
|
48 |
+
"sae_cfg_dict": {
|
49 |
+
"model_name": "gemma-2-2b",
|
50 |
+
"d_in": 2304,
|
51 |
+
"d_sae": 16384,
|
52 |
+
"hook_layer": 12,
|
53 |
+
"hook_name": "blocks.12.hook_resid_post",
|
54 |
+
"context_size": null,
|
55 |
+
"hook_head_index": null,
|
56 |
+
"architecture": "matryoshka_batch_topk",
|
57 |
+
"apply_b_dec_to_input": null,
|
58 |
+
"finetuning_scaling_factor": null,
|
59 |
+
"activation_fn_str": "",
|
60 |
+
"prepend_bos": true,
|
61 |
+
"normalize_activations": "none",
|
62 |
+
"dtype": "bfloat16",
|
63 |
+
"device": "",
|
64 |
+
"dataset_path": "",
|
65 |
+
"dataset_trust_remote_code": true,
|
66 |
+
"seqpos_slice": [
|
67 |
+
null
|
68 |
+
],
|
69 |
+
"training_tokens": -100000,
|
70 |
+
"sae_lens_training_version": null,
|
71 |
+
"neuronpedia_id": null
|
72 |
+
},
|
73 |
+
"eval_result_unstructured": null
|
74 |
+
}
|
unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "unlearning",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"wmdp-bio",
|
7 |
+
"high_school_us_history",
|
8 |
+
"college_computer_science",
|
9 |
+
"high_school_geography",
|
10 |
+
"human_aging"
|
11 |
+
],
|
12 |
+
"intervention_method": "clamp_feature_activation",
|
13 |
+
"retain_thresholds": [
|
14 |
+
0.001,
|
15 |
+
0.01
|
16 |
+
],
|
17 |
+
"n_features_list": [
|
18 |
+
10,
|
19 |
+
20
|
20 |
+
],
|
21 |
+
"multipliers": [
|
22 |
+
25,
|
23 |
+
50,
|
24 |
+
100,
|
25 |
+
200
|
26 |
+
],
|
27 |
+
"dataset_size": 1024,
|
28 |
+
"seq_len": 1024,
|
29 |
+
"n_batch_loss_added": 50,
|
30 |
+
"target_metric": "correct",
|
31 |
+
"save_metrics": true,
|
32 |
+
"model_name": "gemma-2-2b-it",
|
33 |
+
"llm_batch_size": 4,
|
34 |
+
"llm_dtype": "bfloat16"
|
35 |
+
},
|
36 |
+
"eval_id": "fa3a7835-b496-493b-8395-be39948c070b",
|
37 |
+
"datetime_epoch_millis": 1737829722947,
|
38 |
+
"eval_result_metrics": {
|
39 |
+
"unlearning": {
|
40 |
+
"unlearning_score": 0.04690432548522949
|
41 |
+
}
|
42 |
+
},
|
43 |
+
"eval_result_details": [],
|
44 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
45 |
+
"sae_lens_id": "custom_sae",
|
46 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_2",
|
47 |
+
"sae_lens_version": "5.3.2",
|
48 |
+
"sae_cfg_dict": {
|
49 |
+
"model_name": "gemma-2-2b",
|
50 |
+
"d_in": 2304,
|
51 |
+
"d_sae": 16384,
|
52 |
+
"hook_layer": 12,
|
53 |
+
"hook_name": "blocks.12.hook_resid_post",
|
54 |
+
"context_size": null,
|
55 |
+
"hook_head_index": null,
|
56 |
+
"architecture": "matryoshka_batch_topk",
|
57 |
+
"apply_b_dec_to_input": null,
|
58 |
+
"finetuning_scaling_factor": null,
|
59 |
+
"activation_fn_str": "",
|
60 |
+
"prepend_bos": true,
|
61 |
+
"normalize_activations": "none",
|
62 |
+
"dtype": "bfloat16",
|
63 |
+
"device": "",
|
64 |
+
"dataset_path": "",
|
65 |
+
"dataset_trust_remote_code": true,
|
66 |
+
"seqpos_slice": [
|
67 |
+
null
|
68 |
+
],
|
69 |
+
"training_tokens": -100000,
|
70 |
+
"sae_lens_training_version": null,
|
71 |
+
"neuronpedia_id": null
|
72 |
+
},
|
73 |
+
"eval_result_unstructured": null
|
74 |
+
}
|
unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "unlearning",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"wmdp-bio",
|
7 |
+
"high_school_us_history",
|
8 |
+
"college_computer_science",
|
9 |
+
"high_school_geography",
|
10 |
+
"human_aging"
|
11 |
+
],
|
12 |
+
"intervention_method": "clamp_feature_activation",
|
13 |
+
"retain_thresholds": [
|
14 |
+
0.001,
|
15 |
+
0.01
|
16 |
+
],
|
17 |
+
"n_features_list": [
|
18 |
+
10,
|
19 |
+
20
|
20 |
+
],
|
21 |
+
"multipliers": [
|
22 |
+
25,
|
23 |
+
50,
|
24 |
+
100,
|
25 |
+
200
|
26 |
+
],
|
27 |
+
"dataset_size": 1024,
|
28 |
+
"seq_len": 1024,
|
29 |
+
"n_batch_loss_added": 50,
|
30 |
+
"target_metric": "correct",
|
31 |
+
"save_metrics": true,
|
32 |
+
"model_name": "gemma-2-2b-it",
|
33 |
+
"llm_batch_size": 4,
|
34 |
+
"llm_dtype": "bfloat16"
|
35 |
+
},
|
36 |
+
"eval_id": "67f4f3d4-ee7d-4183-8fd0-6458e9bd5c5a",
|
37 |
+
"datetime_epoch_millis": 1737830409224,
|
38 |
+
"eval_result_metrics": {
|
39 |
+
"unlearning": {
|
40 |
+
"unlearning_score": 0.052532851696014404
|
41 |
+
}
|
42 |
+
},
|
43 |
+
"eval_result_details": [],
|
44 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
45 |
+
"sae_lens_id": "custom_sae",
|
46 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_3",
|
47 |
+
"sae_lens_version": "5.3.2",
|
48 |
+
"sae_cfg_dict": {
|
49 |
+
"model_name": "gemma-2-2b",
|
50 |
+
"d_in": 2304,
|
51 |
+
"d_sae": 16384,
|
52 |
+
"hook_layer": 12,
|
53 |
+
"hook_name": "blocks.12.hook_resid_post",
|
54 |
+
"context_size": null,
|
55 |
+
"hook_head_index": null,
|
56 |
+
"architecture": "matryoshka_batch_topk",
|
57 |
+
"apply_b_dec_to_input": null,
|
58 |
+
"finetuning_scaling_factor": null,
|
59 |
+
"activation_fn_str": "",
|
60 |
+
"prepend_bos": true,
|
61 |
+
"normalize_activations": "none",
|
62 |
+
"dtype": "bfloat16",
|
63 |
+
"device": "",
|
64 |
+
"dataset_path": "",
|
65 |
+
"dataset_trust_remote_code": true,
|
66 |
+
"seqpos_slice": [
|
67 |
+
null
|
68 |
+
],
|
69 |
+
"training_tokens": -100000,
|
70 |
+
"sae_lens_training_version": null,
|
71 |
+
"neuronpedia_id": null
|
72 |
+
},
|
73 |
+
"eval_result_unstructured": null
|
74 |
+
}
|
unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "unlearning",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"wmdp-bio",
|
7 |
+
"high_school_us_history",
|
8 |
+
"college_computer_science",
|
9 |
+
"high_school_geography",
|
10 |
+
"human_aging"
|
11 |
+
],
|
12 |
+
"intervention_method": "clamp_feature_activation",
|
13 |
+
"retain_thresholds": [
|
14 |
+
0.001,
|
15 |
+
0.01
|
16 |
+
],
|
17 |
+
"n_features_list": [
|
18 |
+
10,
|
19 |
+
20
|
20 |
+
],
|
21 |
+
"multipliers": [
|
22 |
+
25,
|
23 |
+
50,
|
24 |
+
100,
|
25 |
+
200
|
26 |
+
],
|
27 |
+
"dataset_size": 1024,
|
28 |
+
"seq_len": 1024,
|
29 |
+
"n_batch_loss_added": 50,
|
30 |
+
"target_metric": "correct",
|
31 |
+
"save_metrics": true,
|
32 |
+
"model_name": "gemma-2-2b-it",
|
33 |
+
"llm_batch_size": 4,
|
34 |
+
"llm_dtype": "bfloat16"
|
35 |
+
},
|
36 |
+
"eval_id": "35638051-89cb-4a3b-b037-64b5d5bd5d39",
|
37 |
+
"datetime_epoch_millis": 1737831091265,
|
38 |
+
"eval_result_metrics": {
|
39 |
+
"unlearning": {
|
40 |
+
"unlearning_score": 0.028142571449279785
|
41 |
+
}
|
42 |
+
},
|
43 |
+
"eval_result_details": [],
|
44 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
45 |
+
"sae_lens_id": "custom_sae",
|
46 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_4",
|
47 |
+
"sae_lens_version": "5.3.2",
|
48 |
+
"sae_cfg_dict": {
|
49 |
+
"model_name": "gemma-2-2b",
|
50 |
+
"d_in": 2304,
|
51 |
+
"d_sae": 16384,
|
52 |
+
"hook_layer": 12,
|
53 |
+
"hook_name": "blocks.12.hook_resid_post",
|
54 |
+
"context_size": null,
|
55 |
+
"hook_head_index": null,
|
56 |
+
"architecture": "matryoshka_batch_topk",
|
57 |
+
"apply_b_dec_to_input": null,
|
58 |
+
"finetuning_scaling_factor": null,
|
59 |
+
"activation_fn_str": "",
|
60 |
+
"prepend_bos": true,
|
61 |
+
"normalize_activations": "none",
|
62 |
+
"dtype": "bfloat16",
|
63 |
+
"device": "",
|
64 |
+
"dataset_path": "",
|
65 |
+
"dataset_trust_remote_code": true,
|
66 |
+
"seqpos_slice": [
|
67 |
+
null
|
68 |
+
],
|
69 |
+
"training_tokens": -100000,
|
70 |
+
"sae_lens_training_version": null,
|
71 |
+
"neuronpedia_id": null
|
72 |
+
},
|
73 |
+
"eval_result_unstructured": null
|
74 |
+
}
|
unlearning/matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5_custom_sae_eval_results.json
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"eval_type_id": "unlearning",
|
3 |
+
"eval_config": {
|
4 |
+
"random_seed": 42,
|
5 |
+
"dataset_names": [
|
6 |
+
"wmdp-bio",
|
7 |
+
"high_school_us_history",
|
8 |
+
"college_computer_science",
|
9 |
+
"high_school_geography",
|
10 |
+
"human_aging"
|
11 |
+
],
|
12 |
+
"intervention_method": "clamp_feature_activation",
|
13 |
+
"retain_thresholds": [
|
14 |
+
0.001,
|
15 |
+
0.01
|
16 |
+
],
|
17 |
+
"n_features_list": [
|
18 |
+
10,
|
19 |
+
20
|
20 |
+
],
|
21 |
+
"multipliers": [
|
22 |
+
25,
|
23 |
+
50,
|
24 |
+
100,
|
25 |
+
200
|
26 |
+
],
|
27 |
+
"dataset_size": 1024,
|
28 |
+
"seq_len": 1024,
|
29 |
+
"n_batch_loss_added": 50,
|
30 |
+
"target_metric": "correct",
|
31 |
+
"save_metrics": true,
|
32 |
+
"model_name": "gemma-2-2b-it",
|
33 |
+
"llm_batch_size": 4,
|
34 |
+
"llm_dtype": "bfloat16"
|
35 |
+
},
|
36 |
+
"eval_id": "3914e080-648b-4a90-930f-096b8e0dd140",
|
37 |
+
"datetime_epoch_millis": 1737831767121,
|
38 |
+
"eval_result_metrics": {
|
39 |
+
"unlearning": {
|
40 |
+
"unlearning_score": 0.030018746852874756
|
41 |
+
}
|
42 |
+
},
|
43 |
+
"eval_result_details": [],
|
44 |
+
"sae_bench_commit_hash": "f2d1d982515d2dee706eb23a1ca459b308988764",
|
45 |
+
"sae_lens_id": "custom_sae",
|
46 |
+
"sae_lens_release_id": "matryoshka_0121_MatryoshkaBatchTopKTrainer_gemma_sixteenths_16k_google_gemma-2-2b_matryoshka_batch_top_k_resid_post_layer_12_trainer_5",
|
47 |
+
"sae_lens_version": "5.3.2",
|
48 |
+
"sae_cfg_dict": {
|
49 |
+
"model_name": "gemma-2-2b",
|
50 |
+
"d_in": 2304,
|
51 |
+
"d_sae": 16384,
|
52 |
+
"hook_layer": 12,
|
53 |
+
"hook_name": "blocks.12.hook_resid_post",
|
54 |
+
"context_size": null,
|
55 |
+
"hook_head_index": null,
|
56 |
+
"architecture": "matryoshka_batch_topk",
|
57 |
+
"apply_b_dec_to_input": null,
|
58 |
+
"finetuning_scaling_factor": null,
|
59 |
+
"activation_fn_str": "",
|
60 |
+
"prepend_bos": true,
|
61 |
+
"normalize_activations": "none",
|
62 |
+
"dtype": "bfloat16",
|
63 |
+
"device": "",
|
64 |
+
"dataset_path": "",
|
65 |
+
"dataset_trust_remote_code": true,
|
66 |
+
"seqpos_slice": [
|
67 |
+
null
|
68 |
+
],
|
69 |
+
"training_tokens": -100000,
|
70 |
+
"sae_lens_training_version": null,
|
71 |
+
"neuronpedia_id": null
|
72 |
+
},
|
73 |
+
"eval_result_unstructured": null
|
74 |
+
}
|