jbloom commited on
Commit
d544291
1 Parent(s): d322eaa

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. benchmark_stats.csv +13 -0
  3. benchmark_stats.html +674 -0
  4. benchmark_stats.png +3 -0
  5. v5_128k_layer_0/cfg.json +1 -0
  6. v5_128k_layer_0/metrics.json +1 -0
  7. v5_128k_layer_0/sae_weights.safetensors +3 -0
  8. v5_128k_layer_0/sparsity.safetensors +3 -0
  9. v5_128k_layer_1/cfg.json +1 -0
  10. v5_128k_layer_1/metrics.json +1 -0
  11. v5_128k_layer_1/sae_weights.safetensors +3 -0
  12. v5_128k_layer_1/sparsity.safetensors +3 -0
  13. v5_128k_layer_10/cfg.json +1 -0
  14. v5_128k_layer_10/metrics.json +1 -0
  15. v5_128k_layer_10/sae_weights.safetensors +3 -0
  16. v5_128k_layer_10/sparsity.safetensors +3 -0
  17. v5_128k_layer_11/cfg.json +1 -0
  18. v5_128k_layer_11/metrics.json +1 -0
  19. v5_128k_layer_11/sae_weights.safetensors +3 -0
  20. v5_128k_layer_11/sparsity.safetensors +3 -0
  21. v5_128k_layer_2/cfg.json +1 -0
  22. v5_128k_layer_2/metrics.json +1 -0
  23. v5_128k_layer_2/sae_weights.safetensors +3 -0
  24. v5_128k_layer_2/sparsity.safetensors +3 -0
  25. v5_128k_layer_3/cfg.json +1 -0
  26. v5_128k_layer_3/metrics.json +1 -0
  27. v5_128k_layer_3/sae_weights.safetensors +3 -0
  28. v5_128k_layer_3/sparsity.safetensors +3 -0
  29. v5_128k_layer_4/cfg.json +1 -0
  30. v5_128k_layer_4/metrics.json +1 -0
  31. v5_128k_layer_4/sae_weights.safetensors +3 -0
  32. v5_128k_layer_4/sparsity.safetensors +3 -0
  33. v5_128k_layer_5/cfg.json +1 -0
  34. v5_128k_layer_5/metrics.json +1 -0
  35. v5_128k_layer_5/sae_weights.safetensors +3 -0
  36. v5_128k_layer_5/sparsity.safetensors +3 -0
  37. v5_128k_layer_6/cfg.json +1 -0
  38. v5_128k_layer_6/metrics.json +1 -0
  39. v5_128k_layer_6/sae_weights.safetensors +3 -0
  40. v5_128k_layer_6/sparsity.safetensors +3 -0
  41. v5_128k_layer_7/cfg.json +1 -0
  42. v5_128k_layer_7/metrics.json +1 -0
  43. v5_128k_layer_7/sae_weights.safetensors +3 -0
  44. v5_128k_layer_7/sparsity.safetensors +3 -0
  45. v5_128k_layer_8/cfg.json +1 -0
  46. v5_128k_layer_8/metrics.json +1 -0
  47. v5_128k_layer_8/sae_weights.safetensors +3 -0
  48. v5_128k_layer_8/sparsity.safetensors +3 -0
  49. v5_128k_layer_9/cfg.json +1 -0
  50. v5_128k_layer_9/metrics.json +1 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ benchmark_stats.png filter=lfs diff=lfs merge=lfs -text
benchmark_stats.csv ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ,version,d_sae,layer,kl_div_with_sae,kl_div_with_ablation,ce_loss_with_sae,ce_loss_without_sae,ce_loss_with_ablation,kl_div_score,ce_loss_score,l2_norm_in,l2_norm_out,l2_ratio,l0,l1,explained_variance,mse,total_tokens_evaluated,filepath
2
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_0/metrics.json,5,128,0,0.0028826396446675062,3.094083309173584,3.6018943786621094,3.599064588546753,6.694648742675781,0.999068338064421,0.9990858623204987,29.933448791503906,29.71919822692871,0.9930059909820557,31.99934959411621,61.198638916015625,0.9775193929672241,13.276110649108887,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_0/metrics.json
3
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_1/metrics.json,5,128,1,0.004241525661200285,0.05105271190404892,3.599820852279663,3.599064588546753,3.6525371074676514,0.9169187002411934,0.9858569645086499,18.973735809326172,18.205364227294922,0.9327648878097534,32.0,84.0601806640625,0.9169172048568726,17.682374954223633,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_1/metrics.json
4
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_2/metrics.json,5,128,2,0.006751821842044592,0.05874736234545708,3.602034091949463,3.599064588546753,3.6459126472473145,0.8850702129852012,0.9366141632102598,49.106536865234375,47.9766845703125,0.912467360496521,31.984375,82.67613983154297,0.981090247631073,28.213790893554688,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_2/metrics.json
5
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_3/metrics.json,5,128,3,0.008531898260116577,0.07059153914451599,3.6077229976654053,3.599064588546753,3.658677577972412,0.8791370982484182,0.8547563341132242,16.98731803894043,15.537837028503418,0.8995180130004883,31.90511131286621,81.88644409179688,0.8274940252304077,38.814979553222656,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_3/metrics.json
6
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_4/metrics.json,5,128,4,0.010455957613885403,0.06332532316446304,3.6118924617767334,3.599064588546753,3.6600804328918457,0.8348850492759414,0.7897616042575971,17.251985549926758,15.434064865112305,0.8794459700584412,31.984375,78.37165832519531,0.780777633190155,50.81820297241211,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_4/metrics.json
7
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_5/metrics.json,5,128,5,0.011617729440331459,0.06850520521402359,3.6092593669891357,3.599064588546753,3.6693859100341797,0.8304109971784566,0.8550257841186103,18.888967514038086,16.669851303100586,0.8737667798995972,31.98372459411621,77.13916015625,0.7653716802597046,71.50291442871094,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_5/metrics.json
8
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_6/metrics.json,5,128,6,0.014023388735949993,0.07569437474012375,3.6142406463623047,3.599064588546753,3.67651629447937,0.814736712152052,0.8040577979166154,21.466564178466797,18.890602111816406,0.8757946491241455,31.99837303161621,74.61764526367188,0.754051685333252,97.22125244140625,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_6/metrics.json
9
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_7/metrics.json,5,128,7,0.014005273580551147,0.08048636466264725,3.6098315715789795,3.599064588546753,3.6727118492126465,0.8259919721899077,0.8538033467249813,25.444438934326172,22.559194564819336,0.8843803405761719,31.99788475036621,72.55130004882812,0.7648060321807861,130.78082275390625,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_7/metrics.json
10
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_8/metrics.json,5,128,8,0.0149507075548172,0.08732372522354126,3.6158981323242188,3.599064588546753,3.6803367137908936,0.828789856175459,0.7928743252757569,30.250225067138672,26.934555053710938,0.8884322047233582,31.99690818786621,72.55912780761719,0.7680301070213318,183.63497924804688,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_8/metrics.json
11
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_9/metrics.json,5,128,9,0.01591905765235424,0.09758877754211426,3.615844488143921,3.599064588546753,3.6962451934814453,0.8368761444369522,0.827332834484366,40.192413330078125,36.57688903808594,0.9057919383049011,32.0,68.9762954711914,0.7820959091186523,267.2160949707031,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_9/metrics.json
12
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_10/metrics.json,5,128,10,0.019077030941843987,0.12674781680107117,3.616905927658081,3.599064588546753,3.708984136581421,0.8494882876619082,0.8376872955691096,81.75682830810547,78.83488464355469,0.9609333872795105,32.0,48.323631286621094,0.819786787033081,443.10888671875,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_10/metrics.json
13
+ OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_11/metrics.json,5,128,11,0.024071572348475456,0.17384076118469238,3.620229721069336,3.599064588546753,3.783318281173706,0.8615309080308198,0.8851304838409143,92.90629577636719,88.29003143310547,0.9321191310882568,32.0,71.25521850585938,0.8613338470458984,635.45947265625,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_11/metrics.json
benchmark_stats.html ADDED
@@ -0,0 +1,674 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <style type="text/css">
2
+ #T_a780d_row0_col2, #T_a780d_row0_col3, #T_a780d_row0_col6, #T_a780d_row0_col16, #T_a780d_row0_col17, #T_a780d_row1_col4, #T_a780d_row1_col5, #T_a780d_row1_col6, #T_a780d_row1_col7, #T_a780d_row1_col17, #T_a780d_row2_col4, #T_a780d_row2_col6, #T_a780d_row2_col7, #T_a780d_row2_col17, #T_a780d_row3_col6, #T_a780d_row3_col10, #T_a780d_row3_col11, #T_a780d_row3_col13, #T_a780d_row3_col17, #T_a780d_row4_col6, #T_a780d_row4_col9, #T_a780d_row4_col10, #T_a780d_row4_col11, #T_a780d_row4_col17, #T_a780d_row5_col6, #T_a780d_row5_col12, #T_a780d_row5_col17, #T_a780d_row6_col6, #T_a780d_row6_col8, #T_a780d_row6_col15, #T_a780d_row6_col17, #T_a780d_row7_col6, #T_a780d_row7_col17, #T_a780d_row8_col6, #T_a780d_row8_col17, #T_a780d_row9_col6, #T_a780d_row9_col17, #T_a780d_row10_col6, #T_a780d_row10_col14, #T_a780d_row10_col17, #T_a780d_row11_col6, #T_a780d_row11_col17 {
3
+ background-color: #440154;
4
+ color: #f1f1f1;
5
+ }
6
+ #T_a780d_row0_col4, #T_a780d_row0_col7, #T_a780d_row0_col8, #T_a780d_row0_col9, #T_a780d_row0_col12, #T_a780d_row1_col13, #T_a780d_row1_col14, #T_a780d_row2_col15, #T_a780d_row9_col13, #T_a780d_row10_col13, #T_a780d_row11_col2, #T_a780d_row11_col3, #T_a780d_row11_col5, #T_a780d_row11_col10, #T_a780d_row11_col11, #T_a780d_row11_col13, #T_a780d_row11_col16 {
7
+ background-color: #fde725;
8
+ color: #000000;
9
+ }
10
+ #T_a780d_row0_col5 {
11
+ background-color: #482576;
12
+ color: #f1f1f1;
13
+ }
14
+ #T_a780d_row0_col10 {
15
+ background-color: #443a83;
16
+ color: #f1f1f1;
17
+ }
18
+ #T_a780d_row0_col11 {
19
+ background-color: #414287;
20
+ color: #f1f1f1;
21
+ }
22
+ #T_a780d_row0_col13 {
23
+ background-color: #fbe723;
24
+ color: #000000;
25
+ }
26
+ #T_a780d_row0_col14 {
27
+ background-color: #2e6f8e;
28
+ color: #f1f1f1;
29
+ }
30
+ #T_a780d_row0_col15, #T_a780d_row6_col13 {
31
+ background-color: #f4e61e;
32
+ color: #000000;
33
+ }
34
+ #T_a780d_row1_col2, #T_a780d_row5_col16 {
35
+ background-color: #482173;
36
+ color: #f1f1f1;
37
+ }
38
+ #T_a780d_row1_col3 {
39
+ background-color: #48186a;
40
+ color: #f1f1f1;
41
+ }
42
+ #T_a780d_row1_col8 {
43
+ background-color: #1e9d89;
44
+ color: #f1f1f1;
45
+ }
46
+ #T_a780d_row1_col9 {
47
+ background-color: #d5e21a;
48
+ color: #000000;
49
+ }
50
+ #T_a780d_row1_col10, #T_a780d_row2_col16, #T_a780d_row5_col10, #T_a780d_row10_col4 {
51
+ background-color: #460a5d;
52
+ color: #f1f1f1;
53
+ }
54
+ #T_a780d_row1_col11 {
55
+ background-color: #470e61;
56
+ color: #f1f1f1;
57
+ }
58
+ #T_a780d_row1_col12 {
59
+ background-color: #218f8d;
60
+ color: #f1f1f1;
61
+ }
62
+ #T_a780d_row1_col15 {
63
+ background-color: #4cc26c;
64
+ color: #000000;
65
+ }
66
+ #T_a780d_row1_col16, #T_a780d_row3_col4, #T_a780d_row3_col7, #T_a780d_row4_col4, #T_a780d_row4_col7, #T_a780d_row5_col4, #T_a780d_row5_col7 {
67
+ background-color: #440256;
68
+ color: #f1f1f1;
69
+ }
70
+ #T_a780d_row2_col2, #T_a780d_row2_col3 {
71
+ background-color: #433e85;
72
+ color: #f1f1f1;
73
+ }
74
+ #T_a780d_row2_col5, #T_a780d_row4_col8 {
75
+ background-color: #482677;
76
+ color: #f1f1f1;
77
+ }
78
+ #T_a780d_row2_col8 {
79
+ background-color: #2c738e;
80
+ color: #f1f1f1;
81
+ }
82
+ #T_a780d_row2_col9 {
83
+ background-color: #44bf70;
84
+ color: #f1f1f1;
85
+ }
86
+ #T_a780d_row2_col10 {
87
+ background-color: #277e8e;
88
+ color: #f1f1f1;
89
+ }
90
+ #T_a780d_row2_col11 {
91
+ background-color: #25838e;
92
+ color: #f1f1f1;
93
+ }
94
+ #T_a780d_row2_col12 {
95
+ background-color: #31668e;
96
+ color: #f1f1f1;
97
+ }
98
+ #T_a780d_row2_col13, #T_a780d_row4_col13 {
99
+ background-color: #90d743;
100
+ color: #000000;
101
+ }
102
+ #T_a780d_row2_col14 {
103
+ background-color: #e7e419;
104
+ color: #000000;
105
+ }
106
+ #T_a780d_row3_col2 {
107
+ background-color: #38588c;
108
+ color: #f1f1f1;
109
+ }
110
+ #T_a780d_row3_col3, #T_a780d_row9_col12 {
111
+ background-color: #39568c;
112
+ color: #f1f1f1;
113
+ }
114
+ #T_a780d_row3_col5 {
115
+ background-color: #2b758e;
116
+ color: #f1f1f1;
117
+ }
118
+ #T_a780d_row3_col8 {
119
+ background-color: #2f6c8e;
120
+ color: #f1f1f1;
121
+ }
122
+ #T_a780d_row3_col9, #T_a780d_row5_col9 {
123
+ background-color: #33628d;
124
+ color: #f1f1f1;
125
+ }
126
+ #T_a780d_row3_col12 {
127
+ background-color: #3f4889;
128
+ color: #f1f1f1;
129
+ }
130
+ #T_a780d_row3_col14 {
131
+ background-color: #d8e219;
132
+ color: #000000;
133
+ }
134
+ #T_a780d_row3_col15 {
135
+ background-color: #32658e;
136
+ color: #f1f1f1;
137
+ }
138
+ #T_a780d_row3_col16, #T_a780d_row11_col4 {
139
+ background-color: #471063;
140
+ color: #f1f1f1;
141
+ }
142
+ #T_a780d_row4_col2 {
143
+ background-color: #2d708e;
144
+ color: #f1f1f1;
145
+ }
146
+ #T_a780d_row4_col3 {
147
+ background-color: #2e6e8e;
148
+ color: #f1f1f1;
149
+ }
150
+ #T_a780d_row4_col5 {
151
+ background-color: #21a685;
152
+ color: #f1f1f1;
153
+ }
154
+ #T_a780d_row4_col12, #T_a780d_row5_col15, #T_a780d_row6_col11, #T_a780d_row7_col15 {
155
+ background-color: #471365;
156
+ color: #f1f1f1;
157
+ }
158
+ #T_a780d_row4_col14 {
159
+ background-color: #95d840;
160
+ color: #000000;
161
+ }
162
+ #T_a780d_row4_col15, #T_a780d_row9_col8 {
163
+ background-color: #472a7a;
164
+ color: #f1f1f1;
165
+ }
166
+ #T_a780d_row4_col16, #T_a780d_row6_col10, #T_a780d_row7_col8, #T_a780d_row8_col15 {
167
+ background-color: #481769;
168
+ color: #f1f1f1;
169
+ }
170
+ #T_a780d_row5_col2, #T_a780d_row11_col9 {
171
+ background-color: #25858e;
172
+ color: #f1f1f1;
173
+ }
174
+ #T_a780d_row5_col3 {
175
+ background-color: #297b8e;
176
+ color: #f1f1f1;
177
+ }
178
+ #T_a780d_row5_col5 {
179
+ background-color: #24878e;
180
+ color: #f1f1f1;
181
+ }
182
+ #T_a780d_row5_col8 {
183
+ background-color: #481f70;
184
+ color: #f1f1f1;
185
+ }
186
+ #T_a780d_row5_col11, #T_a780d_row6_col12, #T_a780d_row9_col7 {
187
+ background-color: #46075a;
188
+ color: #f1f1f1;
189
+ }
190
+ #T_a780d_row5_col13 {
191
+ background-color: #8ed645;
192
+ color: #000000;
193
+ }
194
+ #T_a780d_row5_col14 {
195
+ background-color: #7fd34e;
196
+ color: #000000;
197
+ }
198
+ #T_a780d_row6_col2 {
199
+ background-color: #1e9b8a;
200
+ color: #f1f1f1;
201
+ }
202
+ #T_a780d_row6_col3, #T_a780d_row7_col3 {
203
+ background-color: #1f968b;
204
+ color: #f1f1f1;
205
+ }
206
+ #T_a780d_row6_col4, #T_a780d_row6_col7, #T_a780d_row7_col4, #T_a780d_row7_col7, #T_a780d_row8_col7 {
207
+ background-color: #450457;
208
+ color: #f1f1f1;
209
+ }
210
+ #T_a780d_row6_col5 {
211
+ background-color: #46c06f;
212
+ color: #f1f1f1;
213
+ }
214
+ #T_a780d_row6_col9 {
215
+ background-color: #481a6c;
216
+ color: #f1f1f1;
217
+ }
218
+ #T_a780d_row6_col14 {
219
+ background-color: #56c667;
220
+ color: #000000;
221
+ }
222
+ #T_a780d_row6_col16 {
223
+ background-color: #472f7d;
224
+ color: #f1f1f1;
225
+ }
226
+ #T_a780d_row7_col2 {
227
+ background-color: #2ab07f;
228
+ color: #f1f1f1;
229
+ }
230
+ #T_a780d_row7_col5, #T_a780d_row11_col12 {
231
+ background-color: #218e8d;
232
+ color: #f1f1f1;
233
+ }
234
+ #T_a780d_row7_col9, #T_a780d_row9_col10 {
235
+ background-color: #34618d;
236
+ color: #f1f1f1;
237
+ }
238
+ #T_a780d_row7_col10 {
239
+ background-color: #482878;
240
+ color: #f1f1f1;
241
+ }
242
+ #T_a780d_row7_col11 {
243
+ background-color: #482475;
244
+ color: #f1f1f1;
245
+ }
246
+ #T_a780d_row7_col12 {
247
+ background-color: #482071;
248
+ color: #f1f1f1;
249
+ }
250
+ #T_a780d_row7_col13 {
251
+ background-color: #f1e51d;
252
+ color: #000000;
253
+ }
254
+ #T_a780d_row7_col14, #T_a780d_row8_col14 {
255
+ background-color: #3aba76;
256
+ color: #f1f1f1;
257
+ }
258
+ #T_a780d_row7_col16, #T_a780d_row10_col8 {
259
+ background-color: #424086;
260
+ color: #f1f1f1;
261
+ }
262
+ #T_a780d_row8_col2 {
263
+ background-color: #52c569;
264
+ color: #000000;
265
+ }
266
+ #T_a780d_row8_col3 {
267
+ background-color: #1fa188;
268
+ color: #f1f1f1;
269
+ }
270
+ #T_a780d_row8_col4, #T_a780d_row8_col9, #T_a780d_row9_col4 {
271
+ background-color: #450559;
272
+ color: #f1f1f1;
273
+ }
274
+ #T_a780d_row8_col5 {
275
+ background-color: #73d056;
276
+ color: #000000;
277
+ }
278
+ #T_a780d_row8_col8 {
279
+ background-color: #481c6e;
280
+ color: #f1f1f1;
281
+ }
282
+ #T_a780d_row8_col10 {
283
+ background-color: #443b84;
284
+ color: #f1f1f1;
285
+ }
286
+ #T_a780d_row8_col11 {
287
+ background-color: #453781;
288
+ color: #f1f1f1;
289
+ }
290
+ #T_a780d_row8_col12, #T_a780d_row9_col15 {
291
+ background-color: #472c7a;
292
+ color: #f1f1f1;
293
+ }
294
+ #T_a780d_row8_col13 {
295
+ background-color: #eae51a;
296
+ color: #000000;
297
+ }
298
+ #T_a780d_row8_col16 {
299
+ background-color: #38598c;
300
+ color: #f1f1f1;
301
+ }
302
+ #T_a780d_row9_col2 {
303
+ background-color: #86d549;
304
+ color: #000000;
305
+ }
306
+ #T_a780d_row9_col3 {
307
+ background-color: #25ac82;
308
+ color: #f1f1f1;
309
+ }
310
+ #T_a780d_row9_col5 {
311
+ background-color: #70cf57;
312
+ color: #000000;
313
+ }
314
+ #T_a780d_row9_col9 {
315
+ background-color: #433d84;
316
+ color: #f1f1f1;
317
+ }
318
+ #T_a780d_row9_col11, #T_a780d_row10_col15 {
319
+ background-color: #365d8d;
320
+ color: #f1f1f1;
321
+ }
322
+ #T_a780d_row9_col14 {
323
+ background-color: #1fa287;
324
+ color: #f1f1f1;
325
+ }
326
+ #T_a780d_row9_col16 {
327
+ background-color: #297a8e;
328
+ color: #f1f1f1;
329
+ }
330
+ #T_a780d_row10_col2 {
331
+ background-color: #c2df23;
332
+ color: #000000;
333
+ }
334
+ #T_a780d_row10_col3 {
335
+ background-color: #65cb5e;
336
+ color: #000000;
337
+ }
338
+ #T_a780d_row10_col5 {
339
+ background-color: #93d741;
340
+ color: #000000;
341
+ }
342
+ #T_a780d_row10_col7 {
343
+ background-color: #46085c;
344
+ color: #f1f1f1;
345
+ }
346
+ #T_a780d_row10_col9 {
347
+ background-color: #3e4c8a;
348
+ color: #f1f1f1;
349
+ }
350
+ #T_a780d_row10_col10 {
351
+ background-color: #9dd93b;
352
+ color: #000000;
353
+ }
354
+ #T_a780d_row10_col11 {
355
+ background-color: #a8db34;
356
+ color: #000000;
357
+ }
358
+ #T_a780d_row10_col12 {
359
+ background-color: #54c568;
360
+ color: #000000;
361
+ }
362
+ #T_a780d_row10_col16 {
363
+ background-color: #3fbc73;
364
+ color: #f1f1f1;
365
+ }
366
+ #T_a780d_row11_col7 {
367
+ background-color: #471164;
368
+ color: #f1f1f1;
369
+ }
370
+ #T_a780d_row11_col8 {
371
+ background-color: #3b528b;
372
+ color: #f1f1f1;
373
+ }
374
+ #T_a780d_row11_col14 {
375
+ background-color: #2db27d;
376
+ color: #f1f1f1;
377
+ }
378
+ #T_a780d_row11_col15 {
379
+ background-color: #23898e;
380
+ color: #f1f1f1;
381
+ }
382
+ </style>
383
+ <table id="T_a780d">
384
+ <thead>
385
+ <tr>
386
+ <th class="blank level0" >&nbsp;</th>
387
+ <th id="T_a780d_level0_col0" class="col_heading level0 col0" >version</th>
388
+ <th id="T_a780d_level0_col1" class="col_heading level0 col1" >d_sae</th>
389
+ <th id="T_a780d_level0_col2" class="col_heading level0 col2" >layer</th>
390
+ <th id="T_a780d_level0_col3" class="col_heading level0 col3" >kl_div_with_sae</th>
391
+ <th id="T_a780d_level0_col4" class="col_heading level0 col4" >kl_div_with_ablation</th>
392
+ <th id="T_a780d_level0_col5" class="col_heading level0 col5" >ce_loss_with_sae</th>
393
+ <th id="T_a780d_level0_col6" class="col_heading level0 col6" >ce_loss_without_sae</th>
394
+ <th id="T_a780d_level0_col7" class="col_heading level0 col7" >ce_loss_with_ablation</th>
395
+ <th id="T_a780d_level0_col8" class="col_heading level0 col8" >kl_div_score</th>
396
+ <th id="T_a780d_level0_col9" class="col_heading level0 col9" >ce_loss_score</th>
397
+ <th id="T_a780d_level0_col10" class="col_heading level0 col10" >l2_norm_in</th>
398
+ <th id="T_a780d_level0_col11" class="col_heading level0 col11" >l2_norm_out</th>
399
+ <th id="T_a780d_level0_col12" class="col_heading level0 col12" >l2_ratio</th>
400
+ <th id="T_a780d_level0_col13" class="col_heading level0 col13" >l0</th>
401
+ <th id="T_a780d_level0_col14" class="col_heading level0 col14" >l1</th>
402
+ <th id="T_a780d_level0_col15" class="col_heading level0 col15" >explained_variance</th>
403
+ <th id="T_a780d_level0_col16" class="col_heading level0 col16" >mse</th>
404
+ <th id="T_a780d_level0_col17" class="col_heading level0 col17" >total_tokens_evaluated</th>
405
+ <th id="T_a780d_level0_col18" class="col_heading level0 col18" >filepath</th>
406
+ </tr>
407
+ </thead>
408
+ <tbody>
409
+ <tr>
410
+ <th id="T_a780d_level0_row0" class="row_heading level0 row0" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_0/metrics.json</th>
411
+ <td id="T_a780d_row0_col0" class="data row0 col0" >5</td>
412
+ <td id="T_a780d_row0_col1" class="data row0 col1" >128</td>
413
+ <td id="T_a780d_row0_col2" class="data row0 col2" >0</td>
414
+ <td id="T_a780d_row0_col3" class="data row0 col3" >0.002883</td>
415
+ <td id="T_a780d_row0_col4" class="data row0 col4" >3.094083</td>
416
+ <td id="T_a780d_row0_col5" class="data row0 col5" >3.601894</td>
417
+ <td id="T_a780d_row0_col6" class="data row0 col6" >3.599065</td>
418
+ <td id="T_a780d_row0_col7" class="data row0 col7" >6.694649</td>
419
+ <td id="T_a780d_row0_col8" class="data row0 col8" >0.999068</td>
420
+ <td id="T_a780d_row0_col9" class="data row0 col9" >0.999086</td>
421
+ <td id="T_a780d_row0_col10" class="data row0 col10" >29.933449</td>
422
+ <td id="T_a780d_row0_col11" class="data row0 col11" >29.719198</td>
423
+ <td id="T_a780d_row0_col12" class="data row0 col12" >0.993006</td>
424
+ <td id="T_a780d_row0_col13" class="data row0 col13" >31.999350</td>
425
+ <td id="T_a780d_row0_col14" class="data row0 col14" >61.198639</td>
426
+ <td id="T_a780d_row0_col15" class="data row0 col15" >0.977519</td>
427
+ <td id="T_a780d_row0_col16" class="data row0 col16" >13.276111</td>
428
+ <td id="T_a780d_row0_col17" class="data row0 col17" >6144.000000</td>
429
+ <td id="T_a780d_row0_col18" class="data row0 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_0/metrics.json</td>
430
+ </tr>
431
+ <tr>
432
+ <th id="T_a780d_level0_row1" class="row_heading level0 row1" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_1/metrics.json</th>
433
+ <td id="T_a780d_row1_col0" class="data row1 col0" >5</td>
434
+ <td id="T_a780d_row1_col1" class="data row1 col1" >128</td>
435
+ <td id="T_a780d_row1_col2" class="data row1 col2" >1</td>
436
+ <td id="T_a780d_row1_col3" class="data row1 col3" >0.004242</td>
437
+ <td id="T_a780d_row1_col4" class="data row1 col4" >0.051053</td>
438
+ <td id="T_a780d_row1_col5" class="data row1 col5" >3.599821</td>
439
+ <td id="T_a780d_row1_col6" class="data row1 col6" >3.599065</td>
440
+ <td id="T_a780d_row1_col7" class="data row1 col7" >3.652537</td>
441
+ <td id="T_a780d_row1_col8" class="data row1 col8" >0.916919</td>
442
+ <td id="T_a780d_row1_col9" class="data row1 col9" >0.985857</td>
443
+ <td id="T_a780d_row1_col10" class="data row1 col10" >18.973736</td>
444
+ <td id="T_a780d_row1_col11" class="data row1 col11" >18.205364</td>
445
+ <td id="T_a780d_row1_col12" class="data row1 col12" >0.932765</td>
446
+ <td id="T_a780d_row1_col13" class="data row1 col13" >32.000000</td>
447
+ <td id="T_a780d_row1_col14" class="data row1 col14" >84.060181</td>
448
+ <td id="T_a780d_row1_col15" class="data row1 col15" >0.916917</td>
449
+ <td id="T_a780d_row1_col16" class="data row1 col16" >17.682375</td>
450
+ <td id="T_a780d_row1_col17" class="data row1 col17" >6144.000000</td>
451
+ <td id="T_a780d_row1_col18" class="data row1 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_1/metrics.json</td>
452
+ </tr>
453
+ <tr>
454
+ <th id="T_a780d_level0_row2" class="row_heading level0 row2" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_2/metrics.json</th>
455
+ <td id="T_a780d_row2_col0" class="data row2 col0" >5</td>
456
+ <td id="T_a780d_row2_col1" class="data row2 col1" >128</td>
457
+ <td id="T_a780d_row2_col2" class="data row2 col2" >2</td>
458
+ <td id="T_a780d_row2_col3" class="data row2 col3" >0.006752</td>
459
+ <td id="T_a780d_row2_col4" class="data row2 col4" >0.058747</td>
460
+ <td id="T_a780d_row2_col5" class="data row2 col5" >3.602034</td>
461
+ <td id="T_a780d_row2_col6" class="data row2 col6" >3.599065</td>
462
+ <td id="T_a780d_row2_col7" class="data row2 col7" >3.645913</td>
463
+ <td id="T_a780d_row2_col8" class="data row2 col8" >0.885070</td>
464
+ <td id="T_a780d_row2_col9" class="data row2 col9" >0.936614</td>
465
+ <td id="T_a780d_row2_col10" class="data row2 col10" >49.106537</td>
466
+ <td id="T_a780d_row2_col11" class="data row2 col11" >47.976685</td>
467
+ <td id="T_a780d_row2_col12" class="data row2 col12" >0.912467</td>
468
+ <td id="T_a780d_row2_col13" class="data row2 col13" >31.984375</td>
469
+ <td id="T_a780d_row2_col14" class="data row2 col14" >82.676140</td>
470
+ <td id="T_a780d_row2_col15" class="data row2 col15" >0.981090</td>
471
+ <td id="T_a780d_row2_col16" class="data row2 col16" >28.213791</td>
472
+ <td id="T_a780d_row2_col17" class="data row2 col17" >6144.000000</td>
473
+ <td id="T_a780d_row2_col18" class="data row2 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_2/metrics.json</td>
474
+ </tr>
475
+ <tr>
476
+ <th id="T_a780d_level0_row3" class="row_heading level0 row3" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_3/metrics.json</th>
477
+ <td id="T_a780d_row3_col0" class="data row3 col0" >5</td>
478
+ <td id="T_a780d_row3_col1" class="data row3 col1" >128</td>
479
+ <td id="T_a780d_row3_col2" class="data row3 col2" >3</td>
480
+ <td id="T_a780d_row3_col3" class="data row3 col3" >0.008532</td>
481
+ <td id="T_a780d_row3_col4" class="data row3 col4" >0.070592</td>
482
+ <td id="T_a780d_row3_col5" class="data row3 col5" >3.607723</td>
483
+ <td id="T_a780d_row3_col6" class="data row3 col6" >3.599065</td>
484
+ <td id="T_a780d_row3_col7" class="data row3 col7" >3.658678</td>
485
+ <td id="T_a780d_row3_col8" class="data row3 col8" >0.879137</td>
486
+ <td id="T_a780d_row3_col9" class="data row3 col9" >0.854756</td>
487
+ <td id="T_a780d_row3_col10" class="data row3 col10" >16.987318</td>
488
+ <td id="T_a780d_row3_col11" class="data row3 col11" >15.537837</td>
489
+ <td id="T_a780d_row3_col12" class="data row3 col12" >0.899518</td>
490
+ <td id="T_a780d_row3_col13" class="data row3 col13" >31.905111</td>
491
+ <td id="T_a780d_row3_col14" class="data row3 col14" >81.886444</td>
492
+ <td id="T_a780d_row3_col15" class="data row3 col15" >0.827494</td>
493
+ <td id="T_a780d_row3_col16" class="data row3 col16" >38.814980</td>
494
+ <td id="T_a780d_row3_col17" class="data row3 col17" >6144.000000</td>
495
+ <td id="T_a780d_row3_col18" class="data row3 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_3/metrics.json</td>
496
+ </tr>
497
+ <tr>
498
+ <th id="T_a780d_level0_row4" class="row_heading level0 row4" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_4/metrics.json</th>
499
+ <td id="T_a780d_row4_col0" class="data row4 col0" >5</td>
500
+ <td id="T_a780d_row4_col1" class="data row4 col1" >128</td>
501
+ <td id="T_a780d_row4_col2" class="data row4 col2" >4</td>
502
+ <td id="T_a780d_row4_col3" class="data row4 col3" >0.010456</td>
503
+ <td id="T_a780d_row4_col4" class="data row4 col4" >0.063325</td>
504
+ <td id="T_a780d_row4_col5" class="data row4 col5" >3.611892</td>
505
+ <td id="T_a780d_row4_col6" class="data row4 col6" >3.599065</td>
506
+ <td id="T_a780d_row4_col7" class="data row4 col7" >3.660080</td>
507
+ <td id="T_a780d_row4_col8" class="data row4 col8" >0.834885</td>
508
+ <td id="T_a780d_row4_col9" class="data row4 col9" >0.789762</td>
509
+ <td id="T_a780d_row4_col10" class="data row4 col10" >17.251986</td>
510
+ <td id="T_a780d_row4_col11" class="data row4 col11" >15.434065</td>
511
+ <td id="T_a780d_row4_col12" class="data row4 col12" >0.879446</td>
512
+ <td id="T_a780d_row4_col13" class="data row4 col13" >31.984375</td>
513
+ <td id="T_a780d_row4_col14" class="data row4 col14" >78.371658</td>
514
+ <td id="T_a780d_row4_col15" class="data row4 col15" >0.780778</td>
515
+ <td id="T_a780d_row4_col16" class="data row4 col16" >50.818203</td>
516
+ <td id="T_a780d_row4_col17" class="data row4 col17" >6144.000000</td>
517
+ <td id="T_a780d_row4_col18" class="data row4 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_4/metrics.json</td>
518
+ </tr>
519
+ <tr>
520
+ <th id="T_a780d_level0_row5" class="row_heading level0 row5" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_5/metrics.json</th>
521
+ <td id="T_a780d_row5_col0" class="data row5 col0" >5</td>
522
+ <td id="T_a780d_row5_col1" class="data row5 col1" >128</td>
523
+ <td id="T_a780d_row5_col2" class="data row5 col2" >5</td>
524
+ <td id="T_a780d_row5_col3" class="data row5 col3" >0.011618</td>
525
+ <td id="T_a780d_row5_col4" class="data row5 col4" >0.068505</td>
526
+ <td id="T_a780d_row5_col5" class="data row5 col5" >3.609259</td>
527
+ <td id="T_a780d_row5_col6" class="data row5 col6" >3.599065</td>
528
+ <td id="T_a780d_row5_col7" class="data row5 col7" >3.669386</td>
529
+ <td id="T_a780d_row5_col8" class="data row5 col8" >0.830411</td>
530
+ <td id="T_a780d_row5_col9" class="data row5 col9" >0.855026</td>
531
+ <td id="T_a780d_row5_col10" class="data row5 col10" >18.888968</td>
532
+ <td id="T_a780d_row5_col11" class="data row5 col11" >16.669851</td>
533
+ <td id="T_a780d_row5_col12" class="data row5 col12" >0.873767</td>
534
+ <td id="T_a780d_row5_col13" class="data row5 col13" >31.983725</td>
535
+ <td id="T_a780d_row5_col14" class="data row5 col14" >77.139160</td>
536
+ <td id="T_a780d_row5_col15" class="data row5 col15" >0.765372</td>
537
+ <td id="T_a780d_row5_col16" class="data row5 col16" >71.502914</td>
538
+ <td id="T_a780d_row5_col17" class="data row5 col17" >6144.000000</td>
539
+ <td id="T_a780d_row5_col18" class="data row5 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_5/metrics.json</td>
540
+ </tr>
541
+ <tr>
542
+ <th id="T_a780d_level0_row6" class="row_heading level0 row6" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_6/metrics.json</th>
543
+ <td id="T_a780d_row6_col0" class="data row6 col0" >5</td>
544
+ <td id="T_a780d_row6_col1" class="data row6 col1" >128</td>
545
+ <td id="T_a780d_row6_col2" class="data row6 col2" >6</td>
546
+ <td id="T_a780d_row6_col3" class="data row6 col3" >0.014023</td>
547
+ <td id="T_a780d_row6_col4" class="data row6 col4" >0.075694</td>
548
+ <td id="T_a780d_row6_col5" class="data row6 col5" >3.614241</td>
549
+ <td id="T_a780d_row6_col6" class="data row6 col6" >3.599065</td>
550
+ <td id="T_a780d_row6_col7" class="data row6 col7" >3.676516</td>
551
+ <td id="T_a780d_row6_col8" class="data row6 col8" >0.814737</td>
552
+ <td id="T_a780d_row6_col9" class="data row6 col9" >0.804058</td>
553
+ <td id="T_a780d_row6_col10" class="data row6 col10" >21.466564</td>
554
+ <td id="T_a780d_row6_col11" class="data row6 col11" >18.890602</td>
555
+ <td id="T_a780d_row6_col12" class="data row6 col12" >0.875795</td>
556
+ <td id="T_a780d_row6_col13" class="data row6 col13" >31.998373</td>
557
+ <td id="T_a780d_row6_col14" class="data row6 col14" >74.617645</td>
558
+ <td id="T_a780d_row6_col15" class="data row6 col15" >0.754052</td>
559
+ <td id="T_a780d_row6_col16" class="data row6 col16" >97.221252</td>
560
+ <td id="T_a780d_row6_col17" class="data row6 col17" >6144.000000</td>
561
+ <td id="T_a780d_row6_col18" class="data row6 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_6/metrics.json</td>
562
+ </tr>
563
+ <tr>
564
+ <th id="T_a780d_level0_row7" class="row_heading level0 row7" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_7/metrics.json</th>
565
+ <td id="T_a780d_row7_col0" class="data row7 col0" >5</td>
566
+ <td id="T_a780d_row7_col1" class="data row7 col1" >128</td>
567
+ <td id="T_a780d_row7_col2" class="data row7 col2" >7</td>
568
+ <td id="T_a780d_row7_col3" class="data row7 col3" >0.014005</td>
569
+ <td id="T_a780d_row7_col4" class="data row7 col4" >0.080486</td>
570
+ <td id="T_a780d_row7_col5" class="data row7 col5" >3.609832</td>
571
+ <td id="T_a780d_row7_col6" class="data row7 col6" >3.599065</td>
572
+ <td id="T_a780d_row7_col7" class="data row7 col7" >3.672712</td>
573
+ <td id="T_a780d_row7_col8" class="data row7 col8" >0.825992</td>
574
+ <td id="T_a780d_row7_col9" class="data row7 col9" >0.853803</td>
575
+ <td id="T_a780d_row7_col10" class="data row7 col10" >25.444439</td>
576
+ <td id="T_a780d_row7_col11" class="data row7 col11" >22.559195</td>
577
+ <td id="T_a780d_row7_col12" class="data row7 col12" >0.884380</td>
578
+ <td id="T_a780d_row7_col13" class="data row7 col13" >31.997885</td>
579
+ <td id="T_a780d_row7_col14" class="data row7 col14" >72.551300</td>
580
+ <td id="T_a780d_row7_col15" class="data row7 col15" >0.764806</td>
581
+ <td id="T_a780d_row7_col16" class="data row7 col16" >130.780823</td>
582
+ <td id="T_a780d_row7_col17" class="data row7 col17" >6144.000000</td>
583
+ <td id="T_a780d_row7_col18" class="data row7 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_7/metrics.json</td>
584
+ </tr>
585
+ <tr>
586
+ <th id="T_a780d_level0_row8" class="row_heading level0 row8" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_8/metrics.json</th>
587
+ <td id="T_a780d_row8_col0" class="data row8 col0" >5</td>
588
+ <td id="T_a780d_row8_col1" class="data row8 col1" >128</td>
589
+ <td id="T_a780d_row8_col2" class="data row8 col2" >8</td>
590
+ <td id="T_a780d_row8_col3" class="data row8 col3" >0.014951</td>
591
+ <td id="T_a780d_row8_col4" class="data row8 col4" >0.087324</td>
592
+ <td id="T_a780d_row8_col5" class="data row8 col5" >3.615898</td>
593
+ <td id="T_a780d_row8_col6" class="data row8 col6" >3.599065</td>
594
+ <td id="T_a780d_row8_col7" class="data row8 col7" >3.680337</td>
595
+ <td id="T_a780d_row8_col8" class="data row8 col8" >0.828790</td>
596
+ <td id="T_a780d_row8_col9" class="data row8 col9" >0.792874</td>
597
+ <td id="T_a780d_row8_col10" class="data row8 col10" >30.250225</td>
598
+ <td id="T_a780d_row8_col11" class="data row8 col11" >26.934555</td>
599
+ <td id="T_a780d_row8_col12" class="data row8 col12" >0.888432</td>
600
+ <td id="T_a780d_row8_col13" class="data row8 col13" >31.996908</td>
601
+ <td id="T_a780d_row8_col14" class="data row8 col14" >72.559128</td>
602
+ <td id="T_a780d_row8_col15" class="data row8 col15" >0.768030</td>
603
+ <td id="T_a780d_row8_col16" class="data row8 col16" >183.634979</td>
604
+ <td id="T_a780d_row8_col17" class="data row8 col17" >6144.000000</td>
605
+ <td id="T_a780d_row8_col18" class="data row8 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_8/metrics.json</td>
606
+ </tr>
607
+ <tr>
608
+ <th id="T_a780d_level0_row9" class="row_heading level0 row9" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_9/metrics.json</th>
609
+ <td id="T_a780d_row9_col0" class="data row9 col0" >5</td>
610
+ <td id="T_a780d_row9_col1" class="data row9 col1" >128</td>
611
+ <td id="T_a780d_row9_col2" class="data row9 col2" >9</td>
612
+ <td id="T_a780d_row9_col3" class="data row9 col3" >0.015919</td>
613
+ <td id="T_a780d_row9_col4" class="data row9 col4" >0.097589</td>
614
+ <td id="T_a780d_row9_col5" class="data row9 col5" >3.615844</td>
615
+ <td id="T_a780d_row9_col6" class="data row9 col6" >3.599065</td>
616
+ <td id="T_a780d_row9_col7" class="data row9 col7" >3.696245</td>
617
+ <td id="T_a780d_row9_col8" class="data row9 col8" >0.836876</td>
618
+ <td id="T_a780d_row9_col9" class="data row9 col9" >0.827333</td>
619
+ <td id="T_a780d_row9_col10" class="data row9 col10" >40.192413</td>
620
+ <td id="T_a780d_row9_col11" class="data row9 col11" >36.576889</td>
621
+ <td id="T_a780d_row9_col12" class="data row9 col12" >0.905792</td>
622
+ <td id="T_a780d_row9_col13" class="data row9 col13" >32.000000</td>
623
+ <td id="T_a780d_row9_col14" class="data row9 col14" >68.976295</td>
624
+ <td id="T_a780d_row9_col15" class="data row9 col15" >0.782096</td>
625
+ <td id="T_a780d_row9_col16" class="data row9 col16" >267.216095</td>
626
+ <td id="T_a780d_row9_col17" class="data row9 col17" >6144.000000</td>
627
+ <td id="T_a780d_row9_col18" class="data row9 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_9/metrics.json</td>
628
+ </tr>
629
+ <tr>
630
+ <th id="T_a780d_level0_row10" class="row_heading level0 row10" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_10/metrics.json</th>
631
+ <td id="T_a780d_row10_col0" class="data row10 col0" >5</td>
632
+ <td id="T_a780d_row10_col1" class="data row10 col1" >128</td>
633
+ <td id="T_a780d_row10_col2" class="data row10 col2" >10</td>
634
+ <td id="T_a780d_row10_col3" class="data row10 col3" >0.019077</td>
635
+ <td id="T_a780d_row10_col4" class="data row10 col4" >0.126748</td>
636
+ <td id="T_a780d_row10_col5" class="data row10 col5" >3.616906</td>
637
+ <td id="T_a780d_row10_col6" class="data row10 col6" >3.599065</td>
638
+ <td id="T_a780d_row10_col7" class="data row10 col7" >3.708984</td>
639
+ <td id="T_a780d_row10_col8" class="data row10 col8" >0.849488</td>
640
+ <td id="T_a780d_row10_col9" class="data row10 col9" >0.837687</td>
641
+ <td id="T_a780d_row10_col10" class="data row10 col10" >81.756828</td>
642
+ <td id="T_a780d_row10_col11" class="data row10 col11" >78.834885</td>
643
+ <td id="T_a780d_row10_col12" class="data row10 col12" >0.960933</td>
644
+ <td id="T_a780d_row10_col13" class="data row10 col13" >32.000000</td>
645
+ <td id="T_a780d_row10_col14" class="data row10 col14" >48.323631</td>
646
+ <td id="T_a780d_row10_col15" class="data row10 col15" >0.819787</td>
647
+ <td id="T_a780d_row10_col16" class="data row10 col16" >443.108887</td>
648
+ <td id="T_a780d_row10_col17" class="data row10 col17" >6144.000000</td>
649
+ <td id="T_a780d_row10_col18" class="data row10 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_10/metrics.json</td>
650
+ </tr>
651
+ <tr>
652
+ <th id="T_a780d_level0_row11" class="row_heading level0 row11" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_11/metrics.json</th>
653
+ <td id="T_a780d_row11_col0" class="data row11 col0" >5</td>
654
+ <td id="T_a780d_row11_col1" class="data row11 col1" >128</td>
655
+ <td id="T_a780d_row11_col2" class="data row11 col2" >11</td>
656
+ <td id="T_a780d_row11_col3" class="data row11 col3" >0.024072</td>
657
+ <td id="T_a780d_row11_col4" class="data row11 col4" >0.173841</td>
658
+ <td id="T_a780d_row11_col5" class="data row11 col5" >3.620230</td>
659
+ <td id="T_a780d_row11_col6" class="data row11 col6" >3.599065</td>
660
+ <td id="T_a780d_row11_col7" class="data row11 col7" >3.783318</td>
661
+ <td id="T_a780d_row11_col8" class="data row11 col8" >0.861531</td>
662
+ <td id="T_a780d_row11_col9" class="data row11 col9" >0.885130</td>
663
+ <td id="T_a780d_row11_col10" class="data row11 col10" >92.906296</td>
664
+ <td id="T_a780d_row11_col11" class="data row11 col11" >88.290031</td>
665
+ <td id="T_a780d_row11_col12" class="data row11 col12" >0.932119</td>
666
+ <td id="T_a780d_row11_col13" class="data row11 col13" >32.000000</td>
667
+ <td id="T_a780d_row11_col14" class="data row11 col14" >71.255219</td>
668
+ <td id="T_a780d_row11_col15" class="data row11 col15" >0.861334</td>
669
+ <td id="T_a780d_row11_col16" class="data row11 col16" >635.459473</td>
670
+ <td id="T_a780d_row11_col17" class="data row11 col17" >6144.000000</td>
671
+ <td id="T_a780d_row11_col18" class="data row11 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_11/metrics.json</td>
672
+ </tr>
673
+ </tbody>
674
+ </table>
benchmark_stats.png ADDED

Git LFS Details

  • SHA256: 72e2aa138e0d070f20d0e705d5fb2290cb6532aa037cc5f000e680b0be05a4ef
  • Pointer size: 132 Bytes
  • Size of remote file: 4.6 MB
v5_128k_layer_0/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.0.hook_mlp_out", "hook_layer": 0, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_0/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.0028826396446675062, "metrics/kl_div_with_ablation": 3.094083309173584, "metrics/ce_loss_with_sae": 3.6018943786621094, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 6.694648742675781, "metrics/kl_div_score": 0.999068338064421, "metrics/ce_loss_score": 0.9990858623204987, "metrics/l2_norm_in": 29.933448791503906, "metrics/l2_norm_out": 29.71919822692871, "metrics/l2_ratio": 0.9930059909820557, "metrics/l0": 31.99934959411621, "metrics/l1": 61.198638916015625, "metrics/explained_variance": 0.9775193929672241, "metrics/mse": 13.276110649108887, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_0/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e313db224266d10978eb24b4c3e5190d0453204822c65075e077702d52a586c
3
+ size 805834048
v5_128k_layer_0/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f5401432bfa99a5fff073491602e63988a3956cad6fad7b6bec93e25a44cddf
3
+ size 524368
v5_128k_layer_1/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.1.hook_mlp_out", "hook_layer": 1, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_1/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.004241525661200285, "metrics/kl_div_with_ablation": 0.05105271190404892, "metrics/ce_loss_with_sae": 3.599820852279663, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6525371074676514, "metrics/kl_div_score": 0.9169187002411934, "metrics/ce_loss_score": 0.9858569645086499, "metrics/l2_norm_in": 18.973735809326172, "metrics/l2_norm_out": 18.205364227294922, "metrics/l2_ratio": 0.9327648878097534, "metrics/l0": 32.0, "metrics/l1": 84.0601806640625, "metrics/explained_variance": 0.9169172048568726, "metrics/mse": 17.682374954223633, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_1/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46095a456b1e23d7fdd8688853bfcc6797bdc67c5aecc76d61662b3512201582
3
+ size 805834048
v5_128k_layer_1/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70d5f5cfc52d74f2bf6737a5d7fd175f425314dcdd34be553e0ed5058812e1ff
3
+ size 524368
v5_128k_layer_10/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.10.hook_mlp_out", "hook_layer": 10, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_10/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.019077030941843987, "metrics/kl_div_with_ablation": 0.12674781680107117, "metrics/ce_loss_with_sae": 3.616905927658081, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.708984136581421, "metrics/kl_div_score": 0.8494882876619082, "metrics/ce_loss_score": 0.8376872955691096, "metrics/l2_norm_in": 81.75682830810547, "metrics/l2_norm_out": 78.83488464355469, "metrics/l2_ratio": 0.9609333872795105, "metrics/l0": 32.0, "metrics/l1": 48.323631286621094, "metrics/explained_variance": 0.819786787033081, "metrics/mse": 443.10888671875, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_10/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4290e02f0d43cf57ce0050987a6ae00af766f43a47c013c77c63a8745b89b43
3
+ size 805834048
v5_128k_layer_10/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e42a413071da022c6c2e5ba5fe561d358348b76a0126d5f313187b5243e1e8b1
3
+ size 524368
v5_128k_layer_11/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.11.hook_mlp_out", "hook_layer": 11, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_11/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.024071572348475456, "metrics/kl_div_with_ablation": 0.17384076118469238, "metrics/ce_loss_with_sae": 3.620229721069336, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.783318281173706, "metrics/kl_div_score": 0.8615309080308198, "metrics/ce_loss_score": 0.8851304838409143, "metrics/l2_norm_in": 92.90629577636719, "metrics/l2_norm_out": 88.29003143310547, "metrics/l2_ratio": 0.9321191310882568, "metrics/l0": 32.0, "metrics/l1": 71.25521850585938, "metrics/explained_variance": 0.8613338470458984, "metrics/mse": 635.45947265625, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_11/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77f326d4427af047f719a0e8328554ca67bfceaddc28627dc7c97f691509ee28
3
+ size 805834048
v5_128k_layer_11/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25f73118e68ddef034d3e396b3689cb81d3a2d8a1352cfa76175d62b64b31c1d
3
+ size 524368
v5_128k_layer_2/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.2.hook_mlp_out", "hook_layer": 2, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_2/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.006751821842044592, "metrics/kl_div_with_ablation": 0.05874736234545708, "metrics/ce_loss_with_sae": 3.602034091949463, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6459126472473145, "metrics/kl_div_score": 0.8850702129852012, "metrics/ce_loss_score": 0.9366141632102598, "metrics/l2_norm_in": 49.106536865234375, "metrics/l2_norm_out": 47.9766845703125, "metrics/l2_ratio": 0.912467360496521, "metrics/l0": 31.984375, "metrics/l1": 82.67613983154297, "metrics/explained_variance": 0.981090247631073, "metrics/mse": 28.213790893554688, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_2/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d54e5970d48359a47600a92f1f073c0b5517d85adb3da218ac6919d76477c5a
3
+ size 805834048
v5_128k_layer_2/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb0cb9d36290181dd877afed97aa7149ae33b6157465a83f7d8a53ffc0fb1cb9
3
+ size 524368
v5_128k_layer_3/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.3.hook_mlp_out", "hook_layer": 3, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_3/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.008531898260116577, "metrics/kl_div_with_ablation": 0.07059153914451599, "metrics/ce_loss_with_sae": 3.6077229976654053, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.658677577972412, "metrics/kl_div_score": 0.8791370982484182, "metrics/ce_loss_score": 0.8547563341132242, "metrics/l2_norm_in": 16.98731803894043, "metrics/l2_norm_out": 15.537837028503418, "metrics/l2_ratio": 0.8995180130004883, "metrics/l0": 31.90511131286621, "metrics/l1": 81.88644409179688, "metrics/explained_variance": 0.8274940252304077, "metrics/mse": 38.814979553222656, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_3/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39589625da53203638c9c34d238ef6c6548db867b6fd9c99d3a46bea77403159
3
+ size 805834048
v5_128k_layer_3/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:539e71813f11254b3c12d743073e0c34052aa3d41a3f5067b6c28e216b0f8024
3
+ size 524368
v5_128k_layer_4/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.4.hook_mlp_out", "hook_layer": 4, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_4/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.010455957613885403, "metrics/kl_div_with_ablation": 0.06332532316446304, "metrics/ce_loss_with_sae": 3.6118924617767334, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6600804328918457, "metrics/kl_div_score": 0.8348850492759414, "metrics/ce_loss_score": 0.7897616042575971, "metrics/l2_norm_in": 17.251985549926758, "metrics/l2_norm_out": 15.434064865112305, "metrics/l2_ratio": 0.8794459700584412, "metrics/l0": 31.984375, "metrics/l1": 78.37165832519531, "metrics/explained_variance": 0.780777633190155, "metrics/mse": 50.81820297241211, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_4/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6487cb7e7a65f9bcfcea94c2214ab1408192e1d1b24690aa5fb2837218e809e
3
+ size 805834048
v5_128k_layer_4/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7ed6462fe0e62b4cacfb2b21d3e3e18668eba46a042911a04f316c7ece41086
3
+ size 524368
v5_128k_layer_5/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.5.hook_mlp_out", "hook_layer": 5, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_5/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.011617729440331459, "metrics/kl_div_with_ablation": 0.06850520521402359, "metrics/ce_loss_with_sae": 3.6092593669891357, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6693859100341797, "metrics/kl_div_score": 0.8304109971784566, "metrics/ce_loss_score": 0.8550257841186103, "metrics/l2_norm_in": 18.888967514038086, "metrics/l2_norm_out": 16.669851303100586, "metrics/l2_ratio": 0.8737667798995972, "metrics/l0": 31.98372459411621, "metrics/l1": 77.13916015625, "metrics/explained_variance": 0.7653716802597046, "metrics/mse": 71.50291442871094, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_5/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14b24eb29046b3f305fadb6868e2afaf4f42d78ab4666a5240470691141cd301
3
+ size 805834048
v5_128k_layer_5/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59744fae94a6eba1e2e064ffda767c2836208cc63554ab1c2e567eff78593263
3
+ size 524368
v5_128k_layer_6/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.6.hook_mlp_out", "hook_layer": 6, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_6/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.014023388735949993, "metrics/kl_div_with_ablation": 0.07569437474012375, "metrics/ce_loss_with_sae": 3.6142406463623047, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.67651629447937, "metrics/kl_div_score": 0.814736712152052, "metrics/ce_loss_score": 0.8040577979166154, "metrics/l2_norm_in": 21.466564178466797, "metrics/l2_norm_out": 18.890602111816406, "metrics/l2_ratio": 0.8757946491241455, "metrics/l0": 31.99837303161621, "metrics/l1": 74.61764526367188, "metrics/explained_variance": 0.754051685333252, "metrics/mse": 97.22125244140625, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_6/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93f5c45db73104697a3c8e8d708639978852ce69fe715da1c073d5de8124baca
3
+ size 805834048
v5_128k_layer_6/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a872de1296a7b9cd559fdc57ebc575f256bd29f2d08fcaea2abeab10751bb58c
3
+ size 524368
v5_128k_layer_7/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.7.hook_mlp_out", "hook_layer": 7, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_7/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.014005273580551147, "metrics/kl_div_with_ablation": 0.08048636466264725, "metrics/ce_loss_with_sae": 3.6098315715789795, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6727118492126465, "metrics/kl_div_score": 0.8259919721899077, "metrics/ce_loss_score": 0.8538033467249813, "metrics/l2_norm_in": 25.444438934326172, "metrics/l2_norm_out": 22.559194564819336, "metrics/l2_ratio": 0.8843803405761719, "metrics/l0": 31.99788475036621, "metrics/l1": 72.55130004882812, "metrics/explained_variance": 0.7648060321807861, "metrics/mse": 130.78082275390625, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_7/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:caeb6361e42d377cadc85c5fb6d1b488bd9dc4db0223133903ee4df02a78c73f
3
+ size 805834048
v5_128k_layer_7/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5efc7e378076c3fc51ac2e5a98e235b2b3a3aae78305d9055847074ee7c9d699
3
+ size 524368
v5_128k_layer_8/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.8.hook_mlp_out", "hook_layer": 8, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_8/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.0149507075548172, "metrics/kl_div_with_ablation": 0.08732372522354126, "metrics/ce_loss_with_sae": 3.6158981323242188, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6803367137908936, "metrics/kl_div_score": 0.828789856175459, "metrics/ce_loss_score": 0.7928743252757569, "metrics/l2_norm_in": 30.250225067138672, "metrics/l2_norm_out": 26.934555053710938, "metrics/l2_ratio": 0.8884322047233582, "metrics/l0": 31.99690818786621, "metrics/l1": 72.55912780761719, "metrics/explained_variance": 0.7680301070213318, "metrics/mse": 183.63497924804688, "metrics/total_tokens_evaluated": 6144}
v5_128k_layer_8/sae_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bae64cc48c181215c8dd40ca14bb1163dc9ae73ab18296d7b49e7ffe84729485
3
+ size 805834048
v5_128k_layer_8/sparsity.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe8f8bee61f208f771efbcdd8188b56e7a0eb74a32775d68c1adee78a27c3a8e
3
+ size 524368
v5_128k_layer_9/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.9.hook_mlp_out", "hook_layer": 9, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
v5_128k_layer_9/metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metrics/kl_div_with_sae": 0.01591905765235424, "metrics/kl_div_with_ablation": 0.09758877754211426, "metrics/ce_loss_with_sae": 3.615844488143921, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6962451934814453, "metrics/kl_div_score": 0.8368761444369522, "metrics/ce_loss_score": 0.827332834484366, "metrics/l2_norm_in": 40.192413330078125, "metrics/l2_norm_out": 36.57688903808594, "metrics/l2_ratio": 0.9057919383049011, "metrics/l0": 32.0, "metrics/l1": 68.9762954711914, "metrics/explained_variance": 0.7820959091186523, "metrics/mse": 267.2160949707031, "metrics/total_tokens_evaluated": 6144}