Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +1 -0
- benchmark_stats.csv +13 -0
- benchmark_stats.html +674 -0
- benchmark_stats.png +3 -0
- v5_128k_layer_0/cfg.json +1 -0
- v5_128k_layer_0/metrics.json +1 -0
- v5_128k_layer_0/sae_weights.safetensors +3 -0
- v5_128k_layer_0/sparsity.safetensors +3 -0
- v5_128k_layer_1/cfg.json +1 -0
- v5_128k_layer_1/metrics.json +1 -0
- v5_128k_layer_1/sae_weights.safetensors +3 -0
- v5_128k_layer_1/sparsity.safetensors +3 -0
- v5_128k_layer_10/cfg.json +1 -0
- v5_128k_layer_10/metrics.json +1 -0
- v5_128k_layer_10/sae_weights.safetensors +3 -0
- v5_128k_layer_10/sparsity.safetensors +3 -0
- v5_128k_layer_11/cfg.json +1 -0
- v5_128k_layer_11/metrics.json +1 -0
- v5_128k_layer_11/sae_weights.safetensors +3 -0
- v5_128k_layer_11/sparsity.safetensors +3 -0
- v5_128k_layer_2/cfg.json +1 -0
- v5_128k_layer_2/metrics.json +1 -0
- v5_128k_layer_2/sae_weights.safetensors +3 -0
- v5_128k_layer_2/sparsity.safetensors +3 -0
- v5_128k_layer_3/cfg.json +1 -0
- v5_128k_layer_3/metrics.json +1 -0
- v5_128k_layer_3/sae_weights.safetensors +3 -0
- v5_128k_layer_3/sparsity.safetensors +3 -0
- v5_128k_layer_4/cfg.json +1 -0
- v5_128k_layer_4/metrics.json +1 -0
- v5_128k_layer_4/sae_weights.safetensors +3 -0
- v5_128k_layer_4/sparsity.safetensors +3 -0
- v5_128k_layer_5/cfg.json +1 -0
- v5_128k_layer_5/metrics.json +1 -0
- v5_128k_layer_5/sae_weights.safetensors +3 -0
- v5_128k_layer_5/sparsity.safetensors +3 -0
- v5_128k_layer_6/cfg.json +1 -0
- v5_128k_layer_6/metrics.json +1 -0
- v5_128k_layer_6/sae_weights.safetensors +3 -0
- v5_128k_layer_6/sparsity.safetensors +3 -0
- v5_128k_layer_7/cfg.json +1 -0
- v5_128k_layer_7/metrics.json +1 -0
- v5_128k_layer_7/sae_weights.safetensors +3 -0
- v5_128k_layer_7/sparsity.safetensors +3 -0
- v5_128k_layer_8/cfg.json +1 -0
- v5_128k_layer_8/metrics.json +1 -0
- v5_128k_layer_8/sae_weights.safetensors +3 -0
- v5_128k_layer_8/sparsity.safetensors +3 -0
- v5_128k_layer_9/cfg.json +1 -0
- v5_128k_layer_9/metrics.json +1 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
benchmark_stats.png filter=lfs diff=lfs merge=lfs -text
|
benchmark_stats.csv
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
,version,d_sae,layer,kl_div_with_sae,kl_div_with_ablation,ce_loss_with_sae,ce_loss_without_sae,ce_loss_with_ablation,kl_div_score,ce_loss_score,l2_norm_in,l2_norm_out,l2_ratio,l0,l1,explained_variance,mse,total_tokens_evaluated,filepath
|
2 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_0/metrics.json,5,128,0,0.0028826396446675062,3.094083309173584,3.6018943786621094,3.599064588546753,6.694648742675781,0.999068338064421,0.9990858623204987,29.933448791503906,29.71919822692871,0.9930059909820557,31.99934959411621,61.198638916015625,0.9775193929672241,13.276110649108887,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_0/metrics.json
|
3 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_1/metrics.json,5,128,1,0.004241525661200285,0.05105271190404892,3.599820852279663,3.599064588546753,3.6525371074676514,0.9169187002411934,0.9858569645086499,18.973735809326172,18.205364227294922,0.9327648878097534,32.0,84.0601806640625,0.9169172048568726,17.682374954223633,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_1/metrics.json
|
4 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_2/metrics.json,5,128,2,0.006751821842044592,0.05874736234545708,3.602034091949463,3.599064588546753,3.6459126472473145,0.8850702129852012,0.9366141632102598,49.106536865234375,47.9766845703125,0.912467360496521,31.984375,82.67613983154297,0.981090247631073,28.213790893554688,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_2/metrics.json
|
5 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_3/metrics.json,5,128,3,0.008531898260116577,0.07059153914451599,3.6077229976654053,3.599064588546753,3.658677577972412,0.8791370982484182,0.8547563341132242,16.98731803894043,15.537837028503418,0.8995180130004883,31.90511131286621,81.88644409179688,0.8274940252304077,38.814979553222656,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_3/metrics.json
|
6 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_4/metrics.json,5,128,4,0.010455957613885403,0.06332532316446304,3.6118924617767334,3.599064588546753,3.6600804328918457,0.8348850492759414,0.7897616042575971,17.251985549926758,15.434064865112305,0.8794459700584412,31.984375,78.37165832519531,0.780777633190155,50.81820297241211,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_4/metrics.json
|
7 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_5/metrics.json,5,128,5,0.011617729440331459,0.06850520521402359,3.6092593669891357,3.599064588546753,3.6693859100341797,0.8304109971784566,0.8550257841186103,18.888967514038086,16.669851303100586,0.8737667798995972,31.98372459411621,77.13916015625,0.7653716802597046,71.50291442871094,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_5/metrics.json
|
8 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_6/metrics.json,5,128,6,0.014023388735949993,0.07569437474012375,3.6142406463623047,3.599064588546753,3.67651629447937,0.814736712152052,0.8040577979166154,21.466564178466797,18.890602111816406,0.8757946491241455,31.99837303161621,74.61764526367188,0.754051685333252,97.22125244140625,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_6/metrics.json
|
9 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_7/metrics.json,5,128,7,0.014005273580551147,0.08048636466264725,3.6098315715789795,3.599064588546753,3.6727118492126465,0.8259919721899077,0.8538033467249813,25.444438934326172,22.559194564819336,0.8843803405761719,31.99788475036621,72.55130004882812,0.7648060321807861,130.78082275390625,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_7/metrics.json
|
10 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_8/metrics.json,5,128,8,0.0149507075548172,0.08732372522354126,3.6158981323242188,3.599064588546753,3.6803367137908936,0.828789856175459,0.7928743252757569,30.250225067138672,26.934555053710938,0.8884322047233582,31.99690818786621,72.55912780761719,0.7680301070213318,183.63497924804688,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_8/metrics.json
|
11 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_9/metrics.json,5,128,9,0.01591905765235424,0.09758877754211426,3.615844488143921,3.599064588546753,3.6962451934814453,0.8368761444369522,0.827332834484366,40.192413330078125,36.57688903808594,0.9057919383049011,32.0,68.9762954711914,0.7820959091186523,267.2160949707031,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_9/metrics.json
|
12 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_10/metrics.json,5,128,10,0.019077030941843987,0.12674781680107117,3.616905927658081,3.599064588546753,3.708984136581421,0.8494882876619082,0.8376872955691096,81.75682830810547,78.83488464355469,0.9609333872795105,32.0,48.323631286621094,0.819786787033081,443.10888671875,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_10/metrics.json
|
13 |
+
OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_11/metrics.json,5,128,11,0.024071572348475456,0.17384076118469238,3.620229721069336,3.599064588546753,3.783318281173706,0.8615309080308198,0.8851304838409143,92.90629577636719,88.29003143310547,0.9321191310882568,32.0,71.25521850585938,0.8613338470458984,635.45947265625,6144.0,OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_11/metrics.json
|
benchmark_stats.html
ADDED
@@ -0,0 +1,674 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<style type="text/css">
|
2 |
+
#T_a780d_row0_col2, #T_a780d_row0_col3, #T_a780d_row0_col6, #T_a780d_row0_col16, #T_a780d_row0_col17, #T_a780d_row1_col4, #T_a780d_row1_col5, #T_a780d_row1_col6, #T_a780d_row1_col7, #T_a780d_row1_col17, #T_a780d_row2_col4, #T_a780d_row2_col6, #T_a780d_row2_col7, #T_a780d_row2_col17, #T_a780d_row3_col6, #T_a780d_row3_col10, #T_a780d_row3_col11, #T_a780d_row3_col13, #T_a780d_row3_col17, #T_a780d_row4_col6, #T_a780d_row4_col9, #T_a780d_row4_col10, #T_a780d_row4_col11, #T_a780d_row4_col17, #T_a780d_row5_col6, #T_a780d_row5_col12, #T_a780d_row5_col17, #T_a780d_row6_col6, #T_a780d_row6_col8, #T_a780d_row6_col15, #T_a780d_row6_col17, #T_a780d_row7_col6, #T_a780d_row7_col17, #T_a780d_row8_col6, #T_a780d_row8_col17, #T_a780d_row9_col6, #T_a780d_row9_col17, #T_a780d_row10_col6, #T_a780d_row10_col14, #T_a780d_row10_col17, #T_a780d_row11_col6, #T_a780d_row11_col17 {
|
3 |
+
background-color: #440154;
|
4 |
+
color: #f1f1f1;
|
5 |
+
}
|
6 |
+
#T_a780d_row0_col4, #T_a780d_row0_col7, #T_a780d_row0_col8, #T_a780d_row0_col9, #T_a780d_row0_col12, #T_a780d_row1_col13, #T_a780d_row1_col14, #T_a780d_row2_col15, #T_a780d_row9_col13, #T_a780d_row10_col13, #T_a780d_row11_col2, #T_a780d_row11_col3, #T_a780d_row11_col5, #T_a780d_row11_col10, #T_a780d_row11_col11, #T_a780d_row11_col13, #T_a780d_row11_col16 {
|
7 |
+
background-color: #fde725;
|
8 |
+
color: #000000;
|
9 |
+
}
|
10 |
+
#T_a780d_row0_col5 {
|
11 |
+
background-color: #482576;
|
12 |
+
color: #f1f1f1;
|
13 |
+
}
|
14 |
+
#T_a780d_row0_col10 {
|
15 |
+
background-color: #443a83;
|
16 |
+
color: #f1f1f1;
|
17 |
+
}
|
18 |
+
#T_a780d_row0_col11 {
|
19 |
+
background-color: #414287;
|
20 |
+
color: #f1f1f1;
|
21 |
+
}
|
22 |
+
#T_a780d_row0_col13 {
|
23 |
+
background-color: #fbe723;
|
24 |
+
color: #000000;
|
25 |
+
}
|
26 |
+
#T_a780d_row0_col14 {
|
27 |
+
background-color: #2e6f8e;
|
28 |
+
color: #f1f1f1;
|
29 |
+
}
|
30 |
+
#T_a780d_row0_col15, #T_a780d_row6_col13 {
|
31 |
+
background-color: #f4e61e;
|
32 |
+
color: #000000;
|
33 |
+
}
|
34 |
+
#T_a780d_row1_col2, #T_a780d_row5_col16 {
|
35 |
+
background-color: #482173;
|
36 |
+
color: #f1f1f1;
|
37 |
+
}
|
38 |
+
#T_a780d_row1_col3 {
|
39 |
+
background-color: #48186a;
|
40 |
+
color: #f1f1f1;
|
41 |
+
}
|
42 |
+
#T_a780d_row1_col8 {
|
43 |
+
background-color: #1e9d89;
|
44 |
+
color: #f1f1f1;
|
45 |
+
}
|
46 |
+
#T_a780d_row1_col9 {
|
47 |
+
background-color: #d5e21a;
|
48 |
+
color: #000000;
|
49 |
+
}
|
50 |
+
#T_a780d_row1_col10, #T_a780d_row2_col16, #T_a780d_row5_col10, #T_a780d_row10_col4 {
|
51 |
+
background-color: #460a5d;
|
52 |
+
color: #f1f1f1;
|
53 |
+
}
|
54 |
+
#T_a780d_row1_col11 {
|
55 |
+
background-color: #470e61;
|
56 |
+
color: #f1f1f1;
|
57 |
+
}
|
58 |
+
#T_a780d_row1_col12 {
|
59 |
+
background-color: #218f8d;
|
60 |
+
color: #f1f1f1;
|
61 |
+
}
|
62 |
+
#T_a780d_row1_col15 {
|
63 |
+
background-color: #4cc26c;
|
64 |
+
color: #000000;
|
65 |
+
}
|
66 |
+
#T_a780d_row1_col16, #T_a780d_row3_col4, #T_a780d_row3_col7, #T_a780d_row4_col4, #T_a780d_row4_col7, #T_a780d_row5_col4, #T_a780d_row5_col7 {
|
67 |
+
background-color: #440256;
|
68 |
+
color: #f1f1f1;
|
69 |
+
}
|
70 |
+
#T_a780d_row2_col2, #T_a780d_row2_col3 {
|
71 |
+
background-color: #433e85;
|
72 |
+
color: #f1f1f1;
|
73 |
+
}
|
74 |
+
#T_a780d_row2_col5, #T_a780d_row4_col8 {
|
75 |
+
background-color: #482677;
|
76 |
+
color: #f1f1f1;
|
77 |
+
}
|
78 |
+
#T_a780d_row2_col8 {
|
79 |
+
background-color: #2c738e;
|
80 |
+
color: #f1f1f1;
|
81 |
+
}
|
82 |
+
#T_a780d_row2_col9 {
|
83 |
+
background-color: #44bf70;
|
84 |
+
color: #f1f1f1;
|
85 |
+
}
|
86 |
+
#T_a780d_row2_col10 {
|
87 |
+
background-color: #277e8e;
|
88 |
+
color: #f1f1f1;
|
89 |
+
}
|
90 |
+
#T_a780d_row2_col11 {
|
91 |
+
background-color: #25838e;
|
92 |
+
color: #f1f1f1;
|
93 |
+
}
|
94 |
+
#T_a780d_row2_col12 {
|
95 |
+
background-color: #31668e;
|
96 |
+
color: #f1f1f1;
|
97 |
+
}
|
98 |
+
#T_a780d_row2_col13, #T_a780d_row4_col13 {
|
99 |
+
background-color: #90d743;
|
100 |
+
color: #000000;
|
101 |
+
}
|
102 |
+
#T_a780d_row2_col14 {
|
103 |
+
background-color: #e7e419;
|
104 |
+
color: #000000;
|
105 |
+
}
|
106 |
+
#T_a780d_row3_col2 {
|
107 |
+
background-color: #38588c;
|
108 |
+
color: #f1f1f1;
|
109 |
+
}
|
110 |
+
#T_a780d_row3_col3, #T_a780d_row9_col12 {
|
111 |
+
background-color: #39568c;
|
112 |
+
color: #f1f1f1;
|
113 |
+
}
|
114 |
+
#T_a780d_row3_col5 {
|
115 |
+
background-color: #2b758e;
|
116 |
+
color: #f1f1f1;
|
117 |
+
}
|
118 |
+
#T_a780d_row3_col8 {
|
119 |
+
background-color: #2f6c8e;
|
120 |
+
color: #f1f1f1;
|
121 |
+
}
|
122 |
+
#T_a780d_row3_col9, #T_a780d_row5_col9 {
|
123 |
+
background-color: #33628d;
|
124 |
+
color: #f1f1f1;
|
125 |
+
}
|
126 |
+
#T_a780d_row3_col12 {
|
127 |
+
background-color: #3f4889;
|
128 |
+
color: #f1f1f1;
|
129 |
+
}
|
130 |
+
#T_a780d_row3_col14 {
|
131 |
+
background-color: #d8e219;
|
132 |
+
color: #000000;
|
133 |
+
}
|
134 |
+
#T_a780d_row3_col15 {
|
135 |
+
background-color: #32658e;
|
136 |
+
color: #f1f1f1;
|
137 |
+
}
|
138 |
+
#T_a780d_row3_col16, #T_a780d_row11_col4 {
|
139 |
+
background-color: #471063;
|
140 |
+
color: #f1f1f1;
|
141 |
+
}
|
142 |
+
#T_a780d_row4_col2 {
|
143 |
+
background-color: #2d708e;
|
144 |
+
color: #f1f1f1;
|
145 |
+
}
|
146 |
+
#T_a780d_row4_col3 {
|
147 |
+
background-color: #2e6e8e;
|
148 |
+
color: #f1f1f1;
|
149 |
+
}
|
150 |
+
#T_a780d_row4_col5 {
|
151 |
+
background-color: #21a685;
|
152 |
+
color: #f1f1f1;
|
153 |
+
}
|
154 |
+
#T_a780d_row4_col12, #T_a780d_row5_col15, #T_a780d_row6_col11, #T_a780d_row7_col15 {
|
155 |
+
background-color: #471365;
|
156 |
+
color: #f1f1f1;
|
157 |
+
}
|
158 |
+
#T_a780d_row4_col14 {
|
159 |
+
background-color: #95d840;
|
160 |
+
color: #000000;
|
161 |
+
}
|
162 |
+
#T_a780d_row4_col15, #T_a780d_row9_col8 {
|
163 |
+
background-color: #472a7a;
|
164 |
+
color: #f1f1f1;
|
165 |
+
}
|
166 |
+
#T_a780d_row4_col16, #T_a780d_row6_col10, #T_a780d_row7_col8, #T_a780d_row8_col15 {
|
167 |
+
background-color: #481769;
|
168 |
+
color: #f1f1f1;
|
169 |
+
}
|
170 |
+
#T_a780d_row5_col2, #T_a780d_row11_col9 {
|
171 |
+
background-color: #25858e;
|
172 |
+
color: #f1f1f1;
|
173 |
+
}
|
174 |
+
#T_a780d_row5_col3 {
|
175 |
+
background-color: #297b8e;
|
176 |
+
color: #f1f1f1;
|
177 |
+
}
|
178 |
+
#T_a780d_row5_col5 {
|
179 |
+
background-color: #24878e;
|
180 |
+
color: #f1f1f1;
|
181 |
+
}
|
182 |
+
#T_a780d_row5_col8 {
|
183 |
+
background-color: #481f70;
|
184 |
+
color: #f1f1f1;
|
185 |
+
}
|
186 |
+
#T_a780d_row5_col11, #T_a780d_row6_col12, #T_a780d_row9_col7 {
|
187 |
+
background-color: #46075a;
|
188 |
+
color: #f1f1f1;
|
189 |
+
}
|
190 |
+
#T_a780d_row5_col13 {
|
191 |
+
background-color: #8ed645;
|
192 |
+
color: #000000;
|
193 |
+
}
|
194 |
+
#T_a780d_row5_col14 {
|
195 |
+
background-color: #7fd34e;
|
196 |
+
color: #000000;
|
197 |
+
}
|
198 |
+
#T_a780d_row6_col2 {
|
199 |
+
background-color: #1e9b8a;
|
200 |
+
color: #f1f1f1;
|
201 |
+
}
|
202 |
+
#T_a780d_row6_col3, #T_a780d_row7_col3 {
|
203 |
+
background-color: #1f968b;
|
204 |
+
color: #f1f1f1;
|
205 |
+
}
|
206 |
+
#T_a780d_row6_col4, #T_a780d_row6_col7, #T_a780d_row7_col4, #T_a780d_row7_col7, #T_a780d_row8_col7 {
|
207 |
+
background-color: #450457;
|
208 |
+
color: #f1f1f1;
|
209 |
+
}
|
210 |
+
#T_a780d_row6_col5 {
|
211 |
+
background-color: #46c06f;
|
212 |
+
color: #f1f1f1;
|
213 |
+
}
|
214 |
+
#T_a780d_row6_col9 {
|
215 |
+
background-color: #481a6c;
|
216 |
+
color: #f1f1f1;
|
217 |
+
}
|
218 |
+
#T_a780d_row6_col14 {
|
219 |
+
background-color: #56c667;
|
220 |
+
color: #000000;
|
221 |
+
}
|
222 |
+
#T_a780d_row6_col16 {
|
223 |
+
background-color: #472f7d;
|
224 |
+
color: #f1f1f1;
|
225 |
+
}
|
226 |
+
#T_a780d_row7_col2 {
|
227 |
+
background-color: #2ab07f;
|
228 |
+
color: #f1f1f1;
|
229 |
+
}
|
230 |
+
#T_a780d_row7_col5, #T_a780d_row11_col12 {
|
231 |
+
background-color: #218e8d;
|
232 |
+
color: #f1f1f1;
|
233 |
+
}
|
234 |
+
#T_a780d_row7_col9, #T_a780d_row9_col10 {
|
235 |
+
background-color: #34618d;
|
236 |
+
color: #f1f1f1;
|
237 |
+
}
|
238 |
+
#T_a780d_row7_col10 {
|
239 |
+
background-color: #482878;
|
240 |
+
color: #f1f1f1;
|
241 |
+
}
|
242 |
+
#T_a780d_row7_col11 {
|
243 |
+
background-color: #482475;
|
244 |
+
color: #f1f1f1;
|
245 |
+
}
|
246 |
+
#T_a780d_row7_col12 {
|
247 |
+
background-color: #482071;
|
248 |
+
color: #f1f1f1;
|
249 |
+
}
|
250 |
+
#T_a780d_row7_col13 {
|
251 |
+
background-color: #f1e51d;
|
252 |
+
color: #000000;
|
253 |
+
}
|
254 |
+
#T_a780d_row7_col14, #T_a780d_row8_col14 {
|
255 |
+
background-color: #3aba76;
|
256 |
+
color: #f1f1f1;
|
257 |
+
}
|
258 |
+
#T_a780d_row7_col16, #T_a780d_row10_col8 {
|
259 |
+
background-color: #424086;
|
260 |
+
color: #f1f1f1;
|
261 |
+
}
|
262 |
+
#T_a780d_row8_col2 {
|
263 |
+
background-color: #52c569;
|
264 |
+
color: #000000;
|
265 |
+
}
|
266 |
+
#T_a780d_row8_col3 {
|
267 |
+
background-color: #1fa188;
|
268 |
+
color: #f1f1f1;
|
269 |
+
}
|
270 |
+
#T_a780d_row8_col4, #T_a780d_row8_col9, #T_a780d_row9_col4 {
|
271 |
+
background-color: #450559;
|
272 |
+
color: #f1f1f1;
|
273 |
+
}
|
274 |
+
#T_a780d_row8_col5 {
|
275 |
+
background-color: #73d056;
|
276 |
+
color: #000000;
|
277 |
+
}
|
278 |
+
#T_a780d_row8_col8 {
|
279 |
+
background-color: #481c6e;
|
280 |
+
color: #f1f1f1;
|
281 |
+
}
|
282 |
+
#T_a780d_row8_col10 {
|
283 |
+
background-color: #443b84;
|
284 |
+
color: #f1f1f1;
|
285 |
+
}
|
286 |
+
#T_a780d_row8_col11 {
|
287 |
+
background-color: #453781;
|
288 |
+
color: #f1f1f1;
|
289 |
+
}
|
290 |
+
#T_a780d_row8_col12, #T_a780d_row9_col15 {
|
291 |
+
background-color: #472c7a;
|
292 |
+
color: #f1f1f1;
|
293 |
+
}
|
294 |
+
#T_a780d_row8_col13 {
|
295 |
+
background-color: #eae51a;
|
296 |
+
color: #000000;
|
297 |
+
}
|
298 |
+
#T_a780d_row8_col16 {
|
299 |
+
background-color: #38598c;
|
300 |
+
color: #f1f1f1;
|
301 |
+
}
|
302 |
+
#T_a780d_row9_col2 {
|
303 |
+
background-color: #86d549;
|
304 |
+
color: #000000;
|
305 |
+
}
|
306 |
+
#T_a780d_row9_col3 {
|
307 |
+
background-color: #25ac82;
|
308 |
+
color: #f1f1f1;
|
309 |
+
}
|
310 |
+
#T_a780d_row9_col5 {
|
311 |
+
background-color: #70cf57;
|
312 |
+
color: #000000;
|
313 |
+
}
|
314 |
+
#T_a780d_row9_col9 {
|
315 |
+
background-color: #433d84;
|
316 |
+
color: #f1f1f1;
|
317 |
+
}
|
318 |
+
#T_a780d_row9_col11, #T_a780d_row10_col15 {
|
319 |
+
background-color: #365d8d;
|
320 |
+
color: #f1f1f1;
|
321 |
+
}
|
322 |
+
#T_a780d_row9_col14 {
|
323 |
+
background-color: #1fa287;
|
324 |
+
color: #f1f1f1;
|
325 |
+
}
|
326 |
+
#T_a780d_row9_col16 {
|
327 |
+
background-color: #297a8e;
|
328 |
+
color: #f1f1f1;
|
329 |
+
}
|
330 |
+
#T_a780d_row10_col2 {
|
331 |
+
background-color: #c2df23;
|
332 |
+
color: #000000;
|
333 |
+
}
|
334 |
+
#T_a780d_row10_col3 {
|
335 |
+
background-color: #65cb5e;
|
336 |
+
color: #000000;
|
337 |
+
}
|
338 |
+
#T_a780d_row10_col5 {
|
339 |
+
background-color: #93d741;
|
340 |
+
color: #000000;
|
341 |
+
}
|
342 |
+
#T_a780d_row10_col7 {
|
343 |
+
background-color: #46085c;
|
344 |
+
color: #f1f1f1;
|
345 |
+
}
|
346 |
+
#T_a780d_row10_col9 {
|
347 |
+
background-color: #3e4c8a;
|
348 |
+
color: #f1f1f1;
|
349 |
+
}
|
350 |
+
#T_a780d_row10_col10 {
|
351 |
+
background-color: #9dd93b;
|
352 |
+
color: #000000;
|
353 |
+
}
|
354 |
+
#T_a780d_row10_col11 {
|
355 |
+
background-color: #a8db34;
|
356 |
+
color: #000000;
|
357 |
+
}
|
358 |
+
#T_a780d_row10_col12 {
|
359 |
+
background-color: #54c568;
|
360 |
+
color: #000000;
|
361 |
+
}
|
362 |
+
#T_a780d_row10_col16 {
|
363 |
+
background-color: #3fbc73;
|
364 |
+
color: #f1f1f1;
|
365 |
+
}
|
366 |
+
#T_a780d_row11_col7 {
|
367 |
+
background-color: #471164;
|
368 |
+
color: #f1f1f1;
|
369 |
+
}
|
370 |
+
#T_a780d_row11_col8 {
|
371 |
+
background-color: #3b528b;
|
372 |
+
color: #f1f1f1;
|
373 |
+
}
|
374 |
+
#T_a780d_row11_col14 {
|
375 |
+
background-color: #2db27d;
|
376 |
+
color: #f1f1f1;
|
377 |
+
}
|
378 |
+
#T_a780d_row11_col15 {
|
379 |
+
background-color: #23898e;
|
380 |
+
color: #f1f1f1;
|
381 |
+
}
|
382 |
+
</style>
|
383 |
+
<table id="T_a780d">
|
384 |
+
<thead>
|
385 |
+
<tr>
|
386 |
+
<th class="blank level0" > </th>
|
387 |
+
<th id="T_a780d_level0_col0" class="col_heading level0 col0" >version</th>
|
388 |
+
<th id="T_a780d_level0_col1" class="col_heading level0 col1" >d_sae</th>
|
389 |
+
<th id="T_a780d_level0_col2" class="col_heading level0 col2" >layer</th>
|
390 |
+
<th id="T_a780d_level0_col3" class="col_heading level0 col3" >kl_div_with_sae</th>
|
391 |
+
<th id="T_a780d_level0_col4" class="col_heading level0 col4" >kl_div_with_ablation</th>
|
392 |
+
<th id="T_a780d_level0_col5" class="col_heading level0 col5" >ce_loss_with_sae</th>
|
393 |
+
<th id="T_a780d_level0_col6" class="col_heading level0 col6" >ce_loss_without_sae</th>
|
394 |
+
<th id="T_a780d_level0_col7" class="col_heading level0 col7" >ce_loss_with_ablation</th>
|
395 |
+
<th id="T_a780d_level0_col8" class="col_heading level0 col8" >kl_div_score</th>
|
396 |
+
<th id="T_a780d_level0_col9" class="col_heading level0 col9" >ce_loss_score</th>
|
397 |
+
<th id="T_a780d_level0_col10" class="col_heading level0 col10" >l2_norm_in</th>
|
398 |
+
<th id="T_a780d_level0_col11" class="col_heading level0 col11" >l2_norm_out</th>
|
399 |
+
<th id="T_a780d_level0_col12" class="col_heading level0 col12" >l2_ratio</th>
|
400 |
+
<th id="T_a780d_level0_col13" class="col_heading level0 col13" >l0</th>
|
401 |
+
<th id="T_a780d_level0_col14" class="col_heading level0 col14" >l1</th>
|
402 |
+
<th id="T_a780d_level0_col15" class="col_heading level0 col15" >explained_variance</th>
|
403 |
+
<th id="T_a780d_level0_col16" class="col_heading level0 col16" >mse</th>
|
404 |
+
<th id="T_a780d_level0_col17" class="col_heading level0 col17" >total_tokens_evaluated</th>
|
405 |
+
<th id="T_a780d_level0_col18" class="col_heading level0 col18" >filepath</th>
|
406 |
+
</tr>
|
407 |
+
</thead>
|
408 |
+
<tbody>
|
409 |
+
<tr>
|
410 |
+
<th id="T_a780d_level0_row0" class="row_heading level0 row0" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_0/metrics.json</th>
|
411 |
+
<td id="T_a780d_row0_col0" class="data row0 col0" >5</td>
|
412 |
+
<td id="T_a780d_row0_col1" class="data row0 col1" >128</td>
|
413 |
+
<td id="T_a780d_row0_col2" class="data row0 col2" >0</td>
|
414 |
+
<td id="T_a780d_row0_col3" class="data row0 col3" >0.002883</td>
|
415 |
+
<td id="T_a780d_row0_col4" class="data row0 col4" >3.094083</td>
|
416 |
+
<td id="T_a780d_row0_col5" class="data row0 col5" >3.601894</td>
|
417 |
+
<td id="T_a780d_row0_col6" class="data row0 col6" >3.599065</td>
|
418 |
+
<td id="T_a780d_row0_col7" class="data row0 col7" >6.694649</td>
|
419 |
+
<td id="T_a780d_row0_col8" class="data row0 col8" >0.999068</td>
|
420 |
+
<td id="T_a780d_row0_col9" class="data row0 col9" >0.999086</td>
|
421 |
+
<td id="T_a780d_row0_col10" class="data row0 col10" >29.933449</td>
|
422 |
+
<td id="T_a780d_row0_col11" class="data row0 col11" >29.719198</td>
|
423 |
+
<td id="T_a780d_row0_col12" class="data row0 col12" >0.993006</td>
|
424 |
+
<td id="T_a780d_row0_col13" class="data row0 col13" >31.999350</td>
|
425 |
+
<td id="T_a780d_row0_col14" class="data row0 col14" >61.198639</td>
|
426 |
+
<td id="T_a780d_row0_col15" class="data row0 col15" >0.977519</td>
|
427 |
+
<td id="T_a780d_row0_col16" class="data row0 col16" >13.276111</td>
|
428 |
+
<td id="T_a780d_row0_col17" class="data row0 col17" >6144.000000</td>
|
429 |
+
<td id="T_a780d_row0_col18" class="data row0 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_0/metrics.json</td>
|
430 |
+
</tr>
|
431 |
+
<tr>
|
432 |
+
<th id="T_a780d_level0_row1" class="row_heading level0 row1" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_1/metrics.json</th>
|
433 |
+
<td id="T_a780d_row1_col0" class="data row1 col0" >5</td>
|
434 |
+
<td id="T_a780d_row1_col1" class="data row1 col1" >128</td>
|
435 |
+
<td id="T_a780d_row1_col2" class="data row1 col2" >1</td>
|
436 |
+
<td id="T_a780d_row1_col3" class="data row1 col3" >0.004242</td>
|
437 |
+
<td id="T_a780d_row1_col4" class="data row1 col4" >0.051053</td>
|
438 |
+
<td id="T_a780d_row1_col5" class="data row1 col5" >3.599821</td>
|
439 |
+
<td id="T_a780d_row1_col6" class="data row1 col6" >3.599065</td>
|
440 |
+
<td id="T_a780d_row1_col7" class="data row1 col7" >3.652537</td>
|
441 |
+
<td id="T_a780d_row1_col8" class="data row1 col8" >0.916919</td>
|
442 |
+
<td id="T_a780d_row1_col9" class="data row1 col9" >0.985857</td>
|
443 |
+
<td id="T_a780d_row1_col10" class="data row1 col10" >18.973736</td>
|
444 |
+
<td id="T_a780d_row1_col11" class="data row1 col11" >18.205364</td>
|
445 |
+
<td id="T_a780d_row1_col12" class="data row1 col12" >0.932765</td>
|
446 |
+
<td id="T_a780d_row1_col13" class="data row1 col13" >32.000000</td>
|
447 |
+
<td id="T_a780d_row1_col14" class="data row1 col14" >84.060181</td>
|
448 |
+
<td id="T_a780d_row1_col15" class="data row1 col15" >0.916917</td>
|
449 |
+
<td id="T_a780d_row1_col16" class="data row1 col16" >17.682375</td>
|
450 |
+
<td id="T_a780d_row1_col17" class="data row1 col17" >6144.000000</td>
|
451 |
+
<td id="T_a780d_row1_col18" class="data row1 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_1/metrics.json</td>
|
452 |
+
</tr>
|
453 |
+
<tr>
|
454 |
+
<th id="T_a780d_level0_row2" class="row_heading level0 row2" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_2/metrics.json</th>
|
455 |
+
<td id="T_a780d_row2_col0" class="data row2 col0" >5</td>
|
456 |
+
<td id="T_a780d_row2_col1" class="data row2 col1" >128</td>
|
457 |
+
<td id="T_a780d_row2_col2" class="data row2 col2" >2</td>
|
458 |
+
<td id="T_a780d_row2_col3" class="data row2 col3" >0.006752</td>
|
459 |
+
<td id="T_a780d_row2_col4" class="data row2 col4" >0.058747</td>
|
460 |
+
<td id="T_a780d_row2_col5" class="data row2 col5" >3.602034</td>
|
461 |
+
<td id="T_a780d_row2_col6" class="data row2 col6" >3.599065</td>
|
462 |
+
<td id="T_a780d_row2_col7" class="data row2 col7" >3.645913</td>
|
463 |
+
<td id="T_a780d_row2_col8" class="data row2 col8" >0.885070</td>
|
464 |
+
<td id="T_a780d_row2_col9" class="data row2 col9" >0.936614</td>
|
465 |
+
<td id="T_a780d_row2_col10" class="data row2 col10" >49.106537</td>
|
466 |
+
<td id="T_a780d_row2_col11" class="data row2 col11" >47.976685</td>
|
467 |
+
<td id="T_a780d_row2_col12" class="data row2 col12" >0.912467</td>
|
468 |
+
<td id="T_a780d_row2_col13" class="data row2 col13" >31.984375</td>
|
469 |
+
<td id="T_a780d_row2_col14" class="data row2 col14" >82.676140</td>
|
470 |
+
<td id="T_a780d_row2_col15" class="data row2 col15" >0.981090</td>
|
471 |
+
<td id="T_a780d_row2_col16" class="data row2 col16" >28.213791</td>
|
472 |
+
<td id="T_a780d_row2_col17" class="data row2 col17" >6144.000000</td>
|
473 |
+
<td id="T_a780d_row2_col18" class="data row2 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_2/metrics.json</td>
|
474 |
+
</tr>
|
475 |
+
<tr>
|
476 |
+
<th id="T_a780d_level0_row3" class="row_heading level0 row3" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_3/metrics.json</th>
|
477 |
+
<td id="T_a780d_row3_col0" class="data row3 col0" >5</td>
|
478 |
+
<td id="T_a780d_row3_col1" class="data row3 col1" >128</td>
|
479 |
+
<td id="T_a780d_row3_col2" class="data row3 col2" >3</td>
|
480 |
+
<td id="T_a780d_row3_col3" class="data row3 col3" >0.008532</td>
|
481 |
+
<td id="T_a780d_row3_col4" class="data row3 col4" >0.070592</td>
|
482 |
+
<td id="T_a780d_row3_col5" class="data row3 col5" >3.607723</td>
|
483 |
+
<td id="T_a780d_row3_col6" class="data row3 col6" >3.599065</td>
|
484 |
+
<td id="T_a780d_row3_col7" class="data row3 col7" >3.658678</td>
|
485 |
+
<td id="T_a780d_row3_col8" class="data row3 col8" >0.879137</td>
|
486 |
+
<td id="T_a780d_row3_col9" class="data row3 col9" >0.854756</td>
|
487 |
+
<td id="T_a780d_row3_col10" class="data row3 col10" >16.987318</td>
|
488 |
+
<td id="T_a780d_row3_col11" class="data row3 col11" >15.537837</td>
|
489 |
+
<td id="T_a780d_row3_col12" class="data row3 col12" >0.899518</td>
|
490 |
+
<td id="T_a780d_row3_col13" class="data row3 col13" >31.905111</td>
|
491 |
+
<td id="T_a780d_row3_col14" class="data row3 col14" >81.886444</td>
|
492 |
+
<td id="T_a780d_row3_col15" class="data row3 col15" >0.827494</td>
|
493 |
+
<td id="T_a780d_row3_col16" class="data row3 col16" >38.814980</td>
|
494 |
+
<td id="T_a780d_row3_col17" class="data row3 col17" >6144.000000</td>
|
495 |
+
<td id="T_a780d_row3_col18" class="data row3 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_3/metrics.json</td>
|
496 |
+
</tr>
|
497 |
+
<tr>
|
498 |
+
<th id="T_a780d_level0_row4" class="row_heading level0 row4" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_4/metrics.json</th>
|
499 |
+
<td id="T_a780d_row4_col0" class="data row4 col0" >5</td>
|
500 |
+
<td id="T_a780d_row4_col1" class="data row4 col1" >128</td>
|
501 |
+
<td id="T_a780d_row4_col2" class="data row4 col2" >4</td>
|
502 |
+
<td id="T_a780d_row4_col3" class="data row4 col3" >0.010456</td>
|
503 |
+
<td id="T_a780d_row4_col4" class="data row4 col4" >0.063325</td>
|
504 |
+
<td id="T_a780d_row4_col5" class="data row4 col5" >3.611892</td>
|
505 |
+
<td id="T_a780d_row4_col6" class="data row4 col6" >3.599065</td>
|
506 |
+
<td id="T_a780d_row4_col7" class="data row4 col7" >3.660080</td>
|
507 |
+
<td id="T_a780d_row4_col8" class="data row4 col8" >0.834885</td>
|
508 |
+
<td id="T_a780d_row4_col9" class="data row4 col9" >0.789762</td>
|
509 |
+
<td id="T_a780d_row4_col10" class="data row4 col10" >17.251986</td>
|
510 |
+
<td id="T_a780d_row4_col11" class="data row4 col11" >15.434065</td>
|
511 |
+
<td id="T_a780d_row4_col12" class="data row4 col12" >0.879446</td>
|
512 |
+
<td id="T_a780d_row4_col13" class="data row4 col13" >31.984375</td>
|
513 |
+
<td id="T_a780d_row4_col14" class="data row4 col14" >78.371658</td>
|
514 |
+
<td id="T_a780d_row4_col15" class="data row4 col15" >0.780778</td>
|
515 |
+
<td id="T_a780d_row4_col16" class="data row4 col16" >50.818203</td>
|
516 |
+
<td id="T_a780d_row4_col17" class="data row4 col17" >6144.000000</td>
|
517 |
+
<td id="T_a780d_row4_col18" class="data row4 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_4/metrics.json</td>
|
518 |
+
</tr>
|
519 |
+
<tr>
|
520 |
+
<th id="T_a780d_level0_row5" class="row_heading level0 row5" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_5/metrics.json</th>
|
521 |
+
<td id="T_a780d_row5_col0" class="data row5 col0" >5</td>
|
522 |
+
<td id="T_a780d_row5_col1" class="data row5 col1" >128</td>
|
523 |
+
<td id="T_a780d_row5_col2" class="data row5 col2" >5</td>
|
524 |
+
<td id="T_a780d_row5_col3" class="data row5 col3" >0.011618</td>
|
525 |
+
<td id="T_a780d_row5_col4" class="data row5 col4" >0.068505</td>
|
526 |
+
<td id="T_a780d_row5_col5" class="data row5 col5" >3.609259</td>
|
527 |
+
<td id="T_a780d_row5_col6" class="data row5 col6" >3.599065</td>
|
528 |
+
<td id="T_a780d_row5_col7" class="data row5 col7" >3.669386</td>
|
529 |
+
<td id="T_a780d_row5_col8" class="data row5 col8" >0.830411</td>
|
530 |
+
<td id="T_a780d_row5_col9" class="data row5 col9" >0.855026</td>
|
531 |
+
<td id="T_a780d_row5_col10" class="data row5 col10" >18.888968</td>
|
532 |
+
<td id="T_a780d_row5_col11" class="data row5 col11" >16.669851</td>
|
533 |
+
<td id="T_a780d_row5_col12" class="data row5 col12" >0.873767</td>
|
534 |
+
<td id="T_a780d_row5_col13" class="data row5 col13" >31.983725</td>
|
535 |
+
<td id="T_a780d_row5_col14" class="data row5 col14" >77.139160</td>
|
536 |
+
<td id="T_a780d_row5_col15" class="data row5 col15" >0.765372</td>
|
537 |
+
<td id="T_a780d_row5_col16" class="data row5 col16" >71.502914</td>
|
538 |
+
<td id="T_a780d_row5_col17" class="data row5 col17" >6144.000000</td>
|
539 |
+
<td id="T_a780d_row5_col18" class="data row5 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_5/metrics.json</td>
|
540 |
+
</tr>
|
541 |
+
<tr>
|
542 |
+
<th id="T_a780d_level0_row6" class="row_heading level0 row6" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_6/metrics.json</th>
|
543 |
+
<td id="T_a780d_row6_col0" class="data row6 col0" >5</td>
|
544 |
+
<td id="T_a780d_row6_col1" class="data row6 col1" >128</td>
|
545 |
+
<td id="T_a780d_row6_col2" class="data row6 col2" >6</td>
|
546 |
+
<td id="T_a780d_row6_col3" class="data row6 col3" >0.014023</td>
|
547 |
+
<td id="T_a780d_row6_col4" class="data row6 col4" >0.075694</td>
|
548 |
+
<td id="T_a780d_row6_col5" class="data row6 col5" >3.614241</td>
|
549 |
+
<td id="T_a780d_row6_col6" class="data row6 col6" >3.599065</td>
|
550 |
+
<td id="T_a780d_row6_col7" class="data row6 col7" >3.676516</td>
|
551 |
+
<td id="T_a780d_row6_col8" class="data row6 col8" >0.814737</td>
|
552 |
+
<td id="T_a780d_row6_col9" class="data row6 col9" >0.804058</td>
|
553 |
+
<td id="T_a780d_row6_col10" class="data row6 col10" >21.466564</td>
|
554 |
+
<td id="T_a780d_row6_col11" class="data row6 col11" >18.890602</td>
|
555 |
+
<td id="T_a780d_row6_col12" class="data row6 col12" >0.875795</td>
|
556 |
+
<td id="T_a780d_row6_col13" class="data row6 col13" >31.998373</td>
|
557 |
+
<td id="T_a780d_row6_col14" class="data row6 col14" >74.617645</td>
|
558 |
+
<td id="T_a780d_row6_col15" class="data row6 col15" >0.754052</td>
|
559 |
+
<td id="T_a780d_row6_col16" class="data row6 col16" >97.221252</td>
|
560 |
+
<td id="T_a780d_row6_col17" class="data row6 col17" >6144.000000</td>
|
561 |
+
<td id="T_a780d_row6_col18" class="data row6 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_6/metrics.json</td>
|
562 |
+
</tr>
|
563 |
+
<tr>
|
564 |
+
<th id="T_a780d_level0_row7" class="row_heading level0 row7" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_7/metrics.json</th>
|
565 |
+
<td id="T_a780d_row7_col0" class="data row7 col0" >5</td>
|
566 |
+
<td id="T_a780d_row7_col1" class="data row7 col1" >128</td>
|
567 |
+
<td id="T_a780d_row7_col2" class="data row7 col2" >7</td>
|
568 |
+
<td id="T_a780d_row7_col3" class="data row7 col3" >0.014005</td>
|
569 |
+
<td id="T_a780d_row7_col4" class="data row7 col4" >0.080486</td>
|
570 |
+
<td id="T_a780d_row7_col5" class="data row7 col5" >3.609832</td>
|
571 |
+
<td id="T_a780d_row7_col6" class="data row7 col6" >3.599065</td>
|
572 |
+
<td id="T_a780d_row7_col7" class="data row7 col7" >3.672712</td>
|
573 |
+
<td id="T_a780d_row7_col8" class="data row7 col8" >0.825992</td>
|
574 |
+
<td id="T_a780d_row7_col9" class="data row7 col9" >0.853803</td>
|
575 |
+
<td id="T_a780d_row7_col10" class="data row7 col10" >25.444439</td>
|
576 |
+
<td id="T_a780d_row7_col11" class="data row7 col11" >22.559195</td>
|
577 |
+
<td id="T_a780d_row7_col12" class="data row7 col12" >0.884380</td>
|
578 |
+
<td id="T_a780d_row7_col13" class="data row7 col13" >31.997885</td>
|
579 |
+
<td id="T_a780d_row7_col14" class="data row7 col14" >72.551300</td>
|
580 |
+
<td id="T_a780d_row7_col15" class="data row7 col15" >0.764806</td>
|
581 |
+
<td id="T_a780d_row7_col16" class="data row7 col16" >130.780823</td>
|
582 |
+
<td id="T_a780d_row7_col17" class="data row7 col17" >6144.000000</td>
|
583 |
+
<td id="T_a780d_row7_col18" class="data row7 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_7/metrics.json</td>
|
584 |
+
</tr>
|
585 |
+
<tr>
|
586 |
+
<th id="T_a780d_level0_row8" class="row_heading level0 row8" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_8/metrics.json</th>
|
587 |
+
<td id="T_a780d_row8_col0" class="data row8 col0" >5</td>
|
588 |
+
<td id="T_a780d_row8_col1" class="data row8 col1" >128</td>
|
589 |
+
<td id="T_a780d_row8_col2" class="data row8 col2" >8</td>
|
590 |
+
<td id="T_a780d_row8_col3" class="data row8 col3" >0.014951</td>
|
591 |
+
<td id="T_a780d_row8_col4" class="data row8 col4" >0.087324</td>
|
592 |
+
<td id="T_a780d_row8_col5" class="data row8 col5" >3.615898</td>
|
593 |
+
<td id="T_a780d_row8_col6" class="data row8 col6" >3.599065</td>
|
594 |
+
<td id="T_a780d_row8_col7" class="data row8 col7" >3.680337</td>
|
595 |
+
<td id="T_a780d_row8_col8" class="data row8 col8" >0.828790</td>
|
596 |
+
<td id="T_a780d_row8_col9" class="data row8 col9" >0.792874</td>
|
597 |
+
<td id="T_a780d_row8_col10" class="data row8 col10" >30.250225</td>
|
598 |
+
<td id="T_a780d_row8_col11" class="data row8 col11" >26.934555</td>
|
599 |
+
<td id="T_a780d_row8_col12" class="data row8 col12" >0.888432</td>
|
600 |
+
<td id="T_a780d_row8_col13" class="data row8 col13" >31.996908</td>
|
601 |
+
<td id="T_a780d_row8_col14" class="data row8 col14" >72.559128</td>
|
602 |
+
<td id="T_a780d_row8_col15" class="data row8 col15" >0.768030</td>
|
603 |
+
<td id="T_a780d_row8_col16" class="data row8 col16" >183.634979</td>
|
604 |
+
<td id="T_a780d_row8_col17" class="data row8 col17" >6144.000000</td>
|
605 |
+
<td id="T_a780d_row8_col18" class="data row8 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_8/metrics.json</td>
|
606 |
+
</tr>
|
607 |
+
<tr>
|
608 |
+
<th id="T_a780d_level0_row9" class="row_heading level0 row9" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_9/metrics.json</th>
|
609 |
+
<td id="T_a780d_row9_col0" class="data row9 col0" >5</td>
|
610 |
+
<td id="T_a780d_row9_col1" class="data row9 col1" >128</td>
|
611 |
+
<td id="T_a780d_row9_col2" class="data row9 col2" >9</td>
|
612 |
+
<td id="T_a780d_row9_col3" class="data row9 col3" >0.015919</td>
|
613 |
+
<td id="T_a780d_row9_col4" class="data row9 col4" >0.097589</td>
|
614 |
+
<td id="T_a780d_row9_col5" class="data row9 col5" >3.615844</td>
|
615 |
+
<td id="T_a780d_row9_col6" class="data row9 col6" >3.599065</td>
|
616 |
+
<td id="T_a780d_row9_col7" class="data row9 col7" >3.696245</td>
|
617 |
+
<td id="T_a780d_row9_col8" class="data row9 col8" >0.836876</td>
|
618 |
+
<td id="T_a780d_row9_col9" class="data row9 col9" >0.827333</td>
|
619 |
+
<td id="T_a780d_row9_col10" class="data row9 col10" >40.192413</td>
|
620 |
+
<td id="T_a780d_row9_col11" class="data row9 col11" >36.576889</td>
|
621 |
+
<td id="T_a780d_row9_col12" class="data row9 col12" >0.905792</td>
|
622 |
+
<td id="T_a780d_row9_col13" class="data row9 col13" >32.000000</td>
|
623 |
+
<td id="T_a780d_row9_col14" class="data row9 col14" >68.976295</td>
|
624 |
+
<td id="T_a780d_row9_col15" class="data row9 col15" >0.782096</td>
|
625 |
+
<td id="T_a780d_row9_col16" class="data row9 col16" >267.216095</td>
|
626 |
+
<td id="T_a780d_row9_col17" class="data row9 col17" >6144.000000</td>
|
627 |
+
<td id="T_a780d_row9_col18" class="data row9 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_9/metrics.json</td>
|
628 |
+
</tr>
|
629 |
+
<tr>
|
630 |
+
<th id="T_a780d_level0_row10" class="row_heading level0 row10" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_10/metrics.json</th>
|
631 |
+
<td id="T_a780d_row10_col0" class="data row10 col0" >5</td>
|
632 |
+
<td id="T_a780d_row10_col1" class="data row10 col1" >128</td>
|
633 |
+
<td id="T_a780d_row10_col2" class="data row10 col2" >10</td>
|
634 |
+
<td id="T_a780d_row10_col3" class="data row10 col3" >0.019077</td>
|
635 |
+
<td id="T_a780d_row10_col4" class="data row10 col4" >0.126748</td>
|
636 |
+
<td id="T_a780d_row10_col5" class="data row10 col5" >3.616906</td>
|
637 |
+
<td id="T_a780d_row10_col6" class="data row10 col6" >3.599065</td>
|
638 |
+
<td id="T_a780d_row10_col7" class="data row10 col7" >3.708984</td>
|
639 |
+
<td id="T_a780d_row10_col8" class="data row10 col8" >0.849488</td>
|
640 |
+
<td id="T_a780d_row10_col9" class="data row10 col9" >0.837687</td>
|
641 |
+
<td id="T_a780d_row10_col10" class="data row10 col10" >81.756828</td>
|
642 |
+
<td id="T_a780d_row10_col11" class="data row10 col11" >78.834885</td>
|
643 |
+
<td id="T_a780d_row10_col12" class="data row10 col12" >0.960933</td>
|
644 |
+
<td id="T_a780d_row10_col13" class="data row10 col13" >32.000000</td>
|
645 |
+
<td id="T_a780d_row10_col14" class="data row10 col14" >48.323631</td>
|
646 |
+
<td id="T_a780d_row10_col15" class="data row10 col15" >0.819787</td>
|
647 |
+
<td id="T_a780d_row10_col16" class="data row10 col16" >443.108887</td>
|
648 |
+
<td id="T_a780d_row10_col17" class="data row10 col17" >6144.000000</td>
|
649 |
+
<td id="T_a780d_row10_col18" class="data row10 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_10/metrics.json</td>
|
650 |
+
</tr>
|
651 |
+
<tr>
|
652 |
+
<th id="T_a780d_level0_row11" class="row_heading level0 row11" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_11/metrics.json</th>
|
653 |
+
<td id="T_a780d_row11_col0" class="data row11 col0" >5</td>
|
654 |
+
<td id="T_a780d_row11_col1" class="data row11 col1" >128</td>
|
655 |
+
<td id="T_a780d_row11_col2" class="data row11 col2" >11</td>
|
656 |
+
<td id="T_a780d_row11_col3" class="data row11 col3" >0.024072</td>
|
657 |
+
<td id="T_a780d_row11_col4" class="data row11 col4" >0.173841</td>
|
658 |
+
<td id="T_a780d_row11_col5" class="data row11 col5" >3.620230</td>
|
659 |
+
<td id="T_a780d_row11_col6" class="data row11 col6" >3.599065</td>
|
660 |
+
<td id="T_a780d_row11_col7" class="data row11 col7" >3.783318</td>
|
661 |
+
<td id="T_a780d_row11_col8" class="data row11 col8" >0.861531</td>
|
662 |
+
<td id="T_a780d_row11_col9" class="data row11 col9" >0.885130</td>
|
663 |
+
<td id="T_a780d_row11_col10" class="data row11 col10" >92.906296</td>
|
664 |
+
<td id="T_a780d_row11_col11" class="data row11 col11" >88.290031</td>
|
665 |
+
<td id="T_a780d_row11_col12" class="data row11 col12" >0.932119</td>
|
666 |
+
<td id="T_a780d_row11_col13" class="data row11 col13" >32.000000</td>
|
667 |
+
<td id="T_a780d_row11_col14" class="data row11 col14" >71.255219</td>
|
668 |
+
<td id="T_a780d_row11_col15" class="data row11 col15" >0.861334</td>
|
669 |
+
<td id="T_a780d_row11_col16" class="data row11 col16" >635.459473</td>
|
670 |
+
<td id="T_a780d_row11_col17" class="data row11 col17" >6144.000000</td>
|
671 |
+
<td id="T_a780d_row11_col18" class="data row11 col18" >OAI_GPT2Small_v5_128k_resid_delta_mlp/v5_128k_layer_11/metrics.json</td>
|
672 |
+
</tr>
|
673 |
+
</tbody>
|
674 |
+
</table>
|
benchmark_stats.png
ADDED
Git LFS Details
|
v5_128k_layer_0/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.0.hook_mlp_out", "hook_layer": 0, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_0/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.0028826396446675062, "metrics/kl_div_with_ablation": 3.094083309173584, "metrics/ce_loss_with_sae": 3.6018943786621094, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 6.694648742675781, "metrics/kl_div_score": 0.999068338064421, "metrics/ce_loss_score": 0.9990858623204987, "metrics/l2_norm_in": 29.933448791503906, "metrics/l2_norm_out": 29.71919822692871, "metrics/l2_ratio": 0.9930059909820557, "metrics/l0": 31.99934959411621, "metrics/l1": 61.198638916015625, "metrics/explained_variance": 0.9775193929672241, "metrics/mse": 13.276110649108887, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_0/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6e313db224266d10978eb24b4c3e5190d0453204822c65075e077702d52a586c
|
3 |
+
size 805834048
|
v5_128k_layer_0/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8f5401432bfa99a5fff073491602e63988a3956cad6fad7b6bec93e25a44cddf
|
3 |
+
size 524368
|
v5_128k_layer_1/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.1.hook_mlp_out", "hook_layer": 1, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_1/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.004241525661200285, "metrics/kl_div_with_ablation": 0.05105271190404892, "metrics/ce_loss_with_sae": 3.599820852279663, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6525371074676514, "metrics/kl_div_score": 0.9169187002411934, "metrics/ce_loss_score": 0.9858569645086499, "metrics/l2_norm_in": 18.973735809326172, "metrics/l2_norm_out": 18.205364227294922, "metrics/l2_ratio": 0.9327648878097534, "metrics/l0": 32.0, "metrics/l1": 84.0601806640625, "metrics/explained_variance": 0.9169172048568726, "metrics/mse": 17.682374954223633, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_1/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46095a456b1e23d7fdd8688853bfcc6797bdc67c5aecc76d61662b3512201582
|
3 |
+
size 805834048
|
v5_128k_layer_1/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:70d5f5cfc52d74f2bf6737a5d7fd175f425314dcdd34be553e0ed5058812e1ff
|
3 |
+
size 524368
|
v5_128k_layer_10/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.10.hook_mlp_out", "hook_layer": 10, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_10/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.019077030941843987, "metrics/kl_div_with_ablation": 0.12674781680107117, "metrics/ce_loss_with_sae": 3.616905927658081, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.708984136581421, "metrics/kl_div_score": 0.8494882876619082, "metrics/ce_loss_score": 0.8376872955691096, "metrics/l2_norm_in": 81.75682830810547, "metrics/l2_norm_out": 78.83488464355469, "metrics/l2_ratio": 0.9609333872795105, "metrics/l0": 32.0, "metrics/l1": 48.323631286621094, "metrics/explained_variance": 0.819786787033081, "metrics/mse": 443.10888671875, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_10/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4290e02f0d43cf57ce0050987a6ae00af766f43a47c013c77c63a8745b89b43
|
3 |
+
size 805834048
|
v5_128k_layer_10/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e42a413071da022c6c2e5ba5fe561d358348b76a0126d5f313187b5243e1e8b1
|
3 |
+
size 524368
|
v5_128k_layer_11/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.11.hook_mlp_out", "hook_layer": 11, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_11/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.024071572348475456, "metrics/kl_div_with_ablation": 0.17384076118469238, "metrics/ce_loss_with_sae": 3.620229721069336, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.783318281173706, "metrics/kl_div_score": 0.8615309080308198, "metrics/ce_loss_score": 0.8851304838409143, "metrics/l2_norm_in": 92.90629577636719, "metrics/l2_norm_out": 88.29003143310547, "metrics/l2_ratio": 0.9321191310882568, "metrics/l0": 32.0, "metrics/l1": 71.25521850585938, "metrics/explained_variance": 0.8613338470458984, "metrics/mse": 635.45947265625, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_11/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:77f326d4427af047f719a0e8328554ca67bfceaddc28627dc7c97f691509ee28
|
3 |
+
size 805834048
|
v5_128k_layer_11/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:25f73118e68ddef034d3e396b3689cb81d3a2d8a1352cfa76175d62b64b31c1d
|
3 |
+
size 524368
|
v5_128k_layer_2/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.2.hook_mlp_out", "hook_layer": 2, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_2/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.006751821842044592, "metrics/kl_div_with_ablation": 0.05874736234545708, "metrics/ce_loss_with_sae": 3.602034091949463, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6459126472473145, "metrics/kl_div_score": 0.8850702129852012, "metrics/ce_loss_score": 0.9366141632102598, "metrics/l2_norm_in": 49.106536865234375, "metrics/l2_norm_out": 47.9766845703125, "metrics/l2_ratio": 0.912467360496521, "metrics/l0": 31.984375, "metrics/l1": 82.67613983154297, "metrics/explained_variance": 0.981090247631073, "metrics/mse": 28.213790893554688, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_2/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7d54e5970d48359a47600a92f1f073c0b5517d85adb3da218ac6919d76477c5a
|
3 |
+
size 805834048
|
v5_128k_layer_2/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cb0cb9d36290181dd877afed97aa7149ae33b6157465a83f7d8a53ffc0fb1cb9
|
3 |
+
size 524368
|
v5_128k_layer_3/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.3.hook_mlp_out", "hook_layer": 3, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_3/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.008531898260116577, "metrics/kl_div_with_ablation": 0.07059153914451599, "metrics/ce_loss_with_sae": 3.6077229976654053, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.658677577972412, "metrics/kl_div_score": 0.8791370982484182, "metrics/ce_loss_score": 0.8547563341132242, "metrics/l2_norm_in": 16.98731803894043, "metrics/l2_norm_out": 15.537837028503418, "metrics/l2_ratio": 0.8995180130004883, "metrics/l0": 31.90511131286621, "metrics/l1": 81.88644409179688, "metrics/explained_variance": 0.8274940252304077, "metrics/mse": 38.814979553222656, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_3/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:39589625da53203638c9c34d238ef6c6548db867b6fd9c99d3a46bea77403159
|
3 |
+
size 805834048
|
v5_128k_layer_3/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:539e71813f11254b3c12d743073e0c34052aa3d41a3f5067b6c28e216b0f8024
|
3 |
+
size 524368
|
v5_128k_layer_4/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.4.hook_mlp_out", "hook_layer": 4, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_4/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.010455957613885403, "metrics/kl_div_with_ablation": 0.06332532316446304, "metrics/ce_loss_with_sae": 3.6118924617767334, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6600804328918457, "metrics/kl_div_score": 0.8348850492759414, "metrics/ce_loss_score": 0.7897616042575971, "metrics/l2_norm_in": 17.251985549926758, "metrics/l2_norm_out": 15.434064865112305, "metrics/l2_ratio": 0.8794459700584412, "metrics/l0": 31.984375, "metrics/l1": 78.37165832519531, "metrics/explained_variance": 0.780777633190155, "metrics/mse": 50.81820297241211, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_4/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b6487cb7e7a65f9bcfcea94c2214ab1408192e1d1b24690aa5fb2837218e809e
|
3 |
+
size 805834048
|
v5_128k_layer_4/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e7ed6462fe0e62b4cacfb2b21d3e3e18668eba46a042911a04f316c7ece41086
|
3 |
+
size 524368
|
v5_128k_layer_5/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.5.hook_mlp_out", "hook_layer": 5, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_5/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.011617729440331459, "metrics/kl_div_with_ablation": 0.06850520521402359, "metrics/ce_loss_with_sae": 3.6092593669891357, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6693859100341797, "metrics/kl_div_score": 0.8304109971784566, "metrics/ce_loss_score": 0.8550257841186103, "metrics/l2_norm_in": 18.888967514038086, "metrics/l2_norm_out": 16.669851303100586, "metrics/l2_ratio": 0.8737667798995972, "metrics/l0": 31.98372459411621, "metrics/l1": 77.13916015625, "metrics/explained_variance": 0.7653716802597046, "metrics/mse": 71.50291442871094, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_5/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:14b24eb29046b3f305fadb6868e2afaf4f42d78ab4666a5240470691141cd301
|
3 |
+
size 805834048
|
v5_128k_layer_5/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:59744fae94a6eba1e2e064ffda767c2836208cc63554ab1c2e567eff78593263
|
3 |
+
size 524368
|
v5_128k_layer_6/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.6.hook_mlp_out", "hook_layer": 6, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_6/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.014023388735949993, "metrics/kl_div_with_ablation": 0.07569437474012375, "metrics/ce_loss_with_sae": 3.6142406463623047, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.67651629447937, "metrics/kl_div_score": 0.814736712152052, "metrics/ce_loss_score": 0.8040577979166154, "metrics/l2_norm_in": 21.466564178466797, "metrics/l2_norm_out": 18.890602111816406, "metrics/l2_ratio": 0.8757946491241455, "metrics/l0": 31.99837303161621, "metrics/l1": 74.61764526367188, "metrics/explained_variance": 0.754051685333252, "metrics/mse": 97.22125244140625, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_6/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:93f5c45db73104697a3c8e8d708639978852ce69fe715da1c073d5de8124baca
|
3 |
+
size 805834048
|
v5_128k_layer_6/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a872de1296a7b9cd559fdc57ebc575f256bd29f2d08fcaea2abeab10751bb58c
|
3 |
+
size 524368
|
v5_128k_layer_7/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.7.hook_mlp_out", "hook_layer": 7, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_7/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.014005273580551147, "metrics/kl_div_with_ablation": 0.08048636466264725, "metrics/ce_loss_with_sae": 3.6098315715789795, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6727118492126465, "metrics/kl_div_score": 0.8259919721899077, "metrics/ce_loss_score": 0.8538033467249813, "metrics/l2_norm_in": 25.444438934326172, "metrics/l2_norm_out": 22.559194564819336, "metrics/l2_ratio": 0.8843803405761719, "metrics/l0": 31.99788475036621, "metrics/l1": 72.55130004882812, "metrics/explained_variance": 0.7648060321807861, "metrics/mse": 130.78082275390625, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_7/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:caeb6361e42d377cadc85c5fb6d1b488bd9dc4db0223133903ee4df02a78c73f
|
3 |
+
size 805834048
|
v5_128k_layer_7/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5efc7e378076c3fc51ac2e5a98e235b2b3a3aae78305d9055847074ee7c9d699
|
3 |
+
size 524368
|
v5_128k_layer_8/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.8.hook_mlp_out", "hook_layer": 8, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_8/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.0149507075548172, "metrics/kl_div_with_ablation": 0.08732372522354126, "metrics/ce_loss_with_sae": 3.6158981323242188, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6803367137908936, "metrics/kl_div_score": 0.828789856175459, "metrics/ce_loss_score": 0.7928743252757569, "metrics/l2_norm_in": 30.250225067138672, "metrics/l2_norm_out": 26.934555053710938, "metrics/l2_ratio": 0.8884322047233582, "metrics/l0": 31.99690818786621, "metrics/l1": 72.55912780761719, "metrics/explained_variance": 0.7680301070213318, "metrics/mse": 183.63497924804688, "metrics/total_tokens_evaluated": 6144}
|
v5_128k_layer_8/sae_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bae64cc48c181215c8dd40ca14bb1163dc9ae73ab18296d7b49e7ffe84729485
|
3 |
+
size 805834048
|
v5_128k_layer_8/sparsity.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe8f8bee61f208f771efbcdd8188b56e7a0eb74a32775d68c1adee78a27c3a8e
|
3 |
+
size 524368
|
v5_128k_layer_9/cfg.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architecture": "standard", "d_in": 768, "d_sae": 131072, "dtype": "torch.float32", "device": "cuda", "model_name": "gpt2-small", "hook_name": "blocks.9.hook_mlp_out", "hook_layer": 9, "hook_head_index": null, "activation_fn_str": "topk", "activation_fn_kwargs": {"k": 32}, "apply_b_dec_to_input": true, "finetuning_scaling_factor": false, "sae_lens_training_version": null, "prepend_bos": false, "dataset_path": "Skylion007/openwebtext", "dataset_trust_remote_code": true, "context_size": 64, "normalize_activations": "layer_norm"}
|
v5_128k_layer_9/metrics.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"metrics/kl_div_with_sae": 0.01591905765235424, "metrics/kl_div_with_ablation": 0.09758877754211426, "metrics/ce_loss_with_sae": 3.615844488143921, "metrics/ce_loss_without_sae": 3.599064588546753, "metrics/ce_loss_with_ablation": 3.6962451934814453, "metrics/kl_div_score": 0.8368761444369522, "metrics/ce_loss_score": 0.827332834484366, "metrics/l2_norm_in": 40.192413330078125, "metrics/l2_norm_out": 36.57688903808594, "metrics/l2_ratio": 0.9057919383049011, "metrics/l0": 32.0, "metrics/l1": 68.9762954711914, "metrics/explained_variance": 0.7820959091186523, "metrics/mse": 267.2160949707031, "metrics/total_tokens_evaluated": 6144}
|