ishangarg183 commited on
Commit
4d90fab
·
verified ·
1 Parent(s): a8ae341

Add assembled activations for llama32-3b-orpo

Browse files
.gitattributes CHANGED
@@ -78,3 +78,5 @@ v1/assembled_activations/llama32-3b-grpo/features/cross_layer_cosine_drift.csv f
78
  v1/assembled_activations/llama32-3b-grpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
79
  v1/assembled_activations/llama32-3b-kto/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
80
  v1/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
 
 
 
78
  v1/assembled_activations/llama32-3b-grpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
79
  v1/assembled_activations/llama32-3b-kto/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
80
  v1/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
81
+ v1/assembled_activations/llama32-3b-orpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
82
+ v1/assembled_activations/llama32-3b-orpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
v1/assembled_activations/llama32-3b-orpo/activations/activations.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08908a1f1bb7cbcc7b051e93b9ecccc6280edfba25abee6105b299ac066fb3fd
3
+ size 4496133925
v1/assembled_activations/llama32-3b-orpo/checkpoints/final.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e96ca89bc60ce7dbc63986165db73f99f795e61404a963cb69008707d5698a3
3
+ size 10873439922
v1/assembled_activations/llama32-3b-orpo/features/counterfactual_scores.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-orpo/features/counterfactual_scores_by_layer.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-orpo/features/cross_layer_cosine_drift.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0631c58b27513255560ce1c280ce66b0f94bc61a5691fe11ca973625baff15d2
3
+ size 26970952
v1/assembled_activations/llama32-3b-orpo/features/decoder_layer_profiles.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:637bfb42c485d102fcab88d7c26234eff6d922cd92cf856c93d6566b1ca7e9d9
3
+ size 12699008
v1/assembled_activations/llama32-3b-orpo/features/feature_activations.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbc5110ea2ddffa0ef856e548a246c0589e2bec9e065dd5179da4460690aff7a
3
+ size 35935153173
v1/assembled_activations/llama32-3b-orpo/metrics/aggregate_metrics.json ADDED
@@ -0,0 +1,346 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crosscoder_kind": "multilayer_sparc",
3
+ "layers": [
4
+ 24,
5
+ 25,
6
+ 26
7
+ ],
8
+ "topk_mode": "model_balanced_layer_agg",
9
+ "topk": 400,
10
+ "expansion_factor": 8,
11
+ "dict_size": 24576,
12
+ "forced_shared_fraction": 0.06,
13
+ "class_counts": {
14
+ "shared_attenuated": 14574,
15
+ "shared_redirected": 9939,
16
+ "aligned_only": 55,
17
+ "shared_intermediate": 5,
18
+ "other": 3
19
+ },
20
+ "multilayer_class_counts": {
21
+ "drifting_or_rotating": 23044,
22
+ "persistent_shared": 1459,
23
+ "localized_aligned_only": 54,
24
+ "mixed_or_ambiguous": 18,
25
+ "persistent_aligned_only": 1
26
+ },
27
+ "classification_thresholds": {
28
+ "rho_base_only": 0.4,
29
+ "rho_aligned_only": 0.7628208654936128,
30
+ "rho_shared_low": 0.5229447270575759,
31
+ "rho_shared_high": 0.7462659203760826
32
+ },
33
+ "threshold_sensitivity": {
34
+ "original": {
35
+ "shared_attenuated": 14574,
36
+ "shared_redirected": 9939,
37
+ "aligned_only": 55,
38
+ "shared_intermediate": 5,
39
+ "other": 3
40
+ },
41
+ "perturbed": {
42
+ "delta_-0.05": {
43
+ "base_only": 0,
44
+ "aligned_only": 48,
45
+ "shared_aligned": 1474,
46
+ "shared_redirected": 22997,
47
+ "shared_intermediate": 9,
48
+ "shared_attenuated": 47,
49
+ "other": 1
50
+ },
51
+ "delta_+0.05": {
52
+ "base_only": 1,
53
+ "aligned_only": 61,
54
+ "shared_aligned": 0,
55
+ "shared_redirected": 156,
56
+ "shared_intermediate": 4,
57
+ "shared_attenuated": 24351,
58
+ "other": 3
59
+ }
60
+ },
61
+ "perturbation": 0.05
62
+ },
63
+ "class_counts_by_layer": {
64
+ "24": {
65
+ "base_only": 14622,
66
+ "aligned_only": 9954
67
+ },
68
+ "25": {
69
+ "base_only": 21025,
70
+ "aligned_only": 3551
71
+ },
72
+ "26": {
73
+ "base_only": 24446,
74
+ "aligned_only": 130
75
+ }
76
+ },
77
+ "feature_sharing_ratio_by_layer": {
78
+ "24": 0.0,
79
+ "25": 0.0,
80
+ "26": 0.0
81
+ },
82
+ "decoder_amplification_by_layer": {
83
+ "24": {
84
+ "median": 1.0913223558375353,
85
+ "p95": 1.41280373870931
86
+ },
87
+ "25": {
88
+ "median": 1.0839082923798187,
89
+ "p95": 1.2986280885980863
90
+ },
91
+ "26": {
92
+ "median": 1.0572289855848696,
93
+ "p95": 1.2351266728725157
94
+ }
95
+ },
96
+ "classification_thresholds_by_layer": {
97
+ "24": {
98
+ "rho_base_only": 0.531619928142286,
99
+ "rho_aligned_only": 0.531619928142286,
100
+ "rho_shared_low": 0.531619928142286,
101
+ "rho_shared_high": 0.531619928142286
102
+ },
103
+ "25": {
104
+ "rho_base_only": 0.5498768040717406,
105
+ "rho_aligned_only": 0.5498768040717406,
106
+ "rho_shared_low": 0.5498768040717406,
107
+ "rho_shared_high": 0.5498768040717406
108
+ },
109
+ "26": {
110
+ "rho_base_only": 0.6043016691501543,
111
+ "rho_aligned_only": 0.6043016691501543,
112
+ "rho_shared_low": 0.6043016691501543,
113
+ "rho_shared_high": 0.6043016691501543
114
+ }
115
+ },
116
+ "threshold_sensitivity_by_layer": {
117
+ "24": {
118
+ "original": {
119
+ "base_only": 14622,
120
+ "aligned_only": 9954
121
+ },
122
+ "perturbed": {
123
+ "delta_-0.05": {
124
+ "base_only": 3301,
125
+ "aligned_only": 1483,
126
+ "shared_aligned": 1425,
127
+ "shared_redirected": 17932,
128
+ "shared_intermediate": 435,
129
+ "shared_attenuated": 0,
130
+ "other": 0
131
+ },
132
+ "delta_+0.05": {
133
+ "base_only": 23093,
134
+ "aligned_only": 1483,
135
+ "shared_aligned": 0,
136
+ "shared_redirected": 0,
137
+ "shared_intermediate": 0,
138
+ "shared_attenuated": 0,
139
+ "other": 0
140
+ }
141
+ },
142
+ "perturbation": 0.05
143
+ },
144
+ "25": {
145
+ "original": {
146
+ "base_only": 21025,
147
+ "aligned_only": 3551
148
+ },
149
+ "perturbed": {
150
+ "delta_-0.05": {
151
+ "base_only": 5113,
152
+ "aligned_only": 130,
153
+ "shared_aligned": 1446,
154
+ "shared_redirected": 17523,
155
+ "shared_intermediate": 364,
156
+ "shared_attenuated": 0,
157
+ "other": 0
158
+ },
159
+ "delta_+0.05": {
160
+ "base_only": 24446,
161
+ "aligned_only": 130,
162
+ "shared_aligned": 0,
163
+ "shared_redirected": 0,
164
+ "shared_intermediate": 0,
165
+ "shared_attenuated": 0,
166
+ "other": 0
167
+ }
168
+ },
169
+ "perturbation": 0.05
170
+ },
171
+ "26": {
172
+ "original": {
173
+ "base_only": 24446,
174
+ "aligned_only": 130
175
+ },
176
+ "perturbed": {
177
+ "delta_-0.05": {
178
+ "base_only": 23507,
179
+ "aligned_only": 91,
180
+ "shared_aligned": 2,
181
+ "shared_redirected": 920,
182
+ "shared_intermediate": 56,
183
+ "shared_attenuated": 0,
184
+ "other": 0
185
+ },
186
+ "delta_+0.05": {
187
+ "base_only": 24485,
188
+ "aligned_only": 91,
189
+ "shared_aligned": 0,
190
+ "shared_redirected": 0,
191
+ "shared_intermediate": 0,
192
+ "shared_attenuated": 0,
193
+ "other": 0
194
+ }
195
+ },
196
+ "perturbation": 0.05
197
+ }
198
+ },
199
+ "counterfactual_shift_by_layer": {
200
+ "24": {
201
+ "aligned_only": {
202
+ "mean_shift": 0.01766282921238268,
203
+ "median_shift": 0.0,
204
+ "p95_abs_shift": 2.36672351360321,
205
+ "count": 55
206
+ },
207
+ "other": {
208
+ "mean_shift": -9.973655323847197e-07,
209
+ "median_shift": 0.0,
210
+ "p95_abs_shift": 0.0,
211
+ "count": 3
212
+ },
213
+ "shared_attenuated": {
214
+ "mean_shift": 0.0006880431172107163,
215
+ "median_shift": 0.0,
216
+ "p95_abs_shift": 0.0,
217
+ "count": 14574
218
+ },
219
+ "shared_intermediate": {
220
+ "mean_shift": 0.0007839000825697439,
221
+ "median_shift": 0.000433885317761451,
222
+ "p95_abs_shift": 0.0,
223
+ "count": 5
224
+ },
225
+ "shared_redirected": {
226
+ "mean_shift": 1.455451122652191e-05,
227
+ "median_shift": 0.0,
228
+ "p95_abs_shift": 0.0,
229
+ "count": 9939
230
+ }
231
+ },
232
+ "25": {
233
+ "aligned_only": {
234
+ "mean_shift": 0.04922456900294492,
235
+ "median_shift": 0.0,
236
+ "p95_abs_shift": 5.535092401504513,
237
+ "count": 55
238
+ },
239
+ "other": {
240
+ "mean_shift": 1.2010831748436128e-06,
241
+ "median_shift": 0.0,
242
+ "p95_abs_shift": 0.0,
243
+ "count": 3
244
+ },
245
+ "shared_attenuated": {
246
+ "mean_shift": 0.00019313933181176285,
247
+ "median_shift": 0.0,
248
+ "p95_abs_shift": 0.0,
249
+ "count": 14574
250
+ },
251
+ "shared_intermediate": {
252
+ "mean_shift": 0.001000284298788756,
253
+ "median_shift": 0.0011236919090151787,
254
+ "p95_abs_shift": 0.0,
255
+ "count": 5
256
+ },
257
+ "shared_redirected": {
258
+ "mean_shift": 1.8357591667978446e-05,
259
+ "median_shift": 0.0,
260
+ "p95_abs_shift": 0.0,
261
+ "count": 9939
262
+ }
263
+ },
264
+ "26": {
265
+ "aligned_only": {
266
+ "mean_shift": 0.02845165677870434,
267
+ "median_shift": 0.0,
268
+ "p95_abs_shift": 3.3389091253280627,
269
+ "count": 55
270
+ },
271
+ "other": {
272
+ "mean_shift": 0.01766254846006632,
273
+ "median_shift": 0.019211646169424057,
274
+ "p95_abs_shift": 0.0,
275
+ "count": 3
276
+ },
277
+ "shared_attenuated": {
278
+ "mean_shift": 9.978226535641974e-05,
279
+ "median_shift": 0.0,
280
+ "p95_abs_shift": 0.0,
281
+ "count": 14574
282
+ },
283
+ "shared_intermediate": {
284
+ "mean_shift": 0.0008990645757876337,
285
+ "median_shift": 0.0004914115997962654,
286
+ "p95_abs_shift": 0.0,
287
+ "count": 5
288
+ },
289
+ "shared_redirected": {
290
+ "mean_shift": 2.211129789160676e-05,
291
+ "median_shift": 0.0,
292
+ "p95_abs_shift": 0.0,
293
+ "count": 9939
294
+ }
295
+ }
296
+ },
297
+ "total_features": 24576,
298
+ "fve_base": 0.6388642060819096,
299
+ "fve_aligned": 0.7197545064681488,
300
+ "fve_base_by_layer": [
301
+ 0.6299517154693604,
302
+ 0.6417424791146323,
303
+ 0.644898364681224
304
+ ],
305
+ "fve_aligned_by_layer": [
306
+ 0.7138387008487243,
307
+ 0.7231695330579867,
308
+ 0.7222552165311045
309
+ ],
310
+ "val_fve_base_by_layer": [
311
+ 0.6299517154693604,
312
+ 0.6417424791146323,
313
+ 0.644898364681224
314
+ ],
315
+ "val_fve_aligned_by_layer": [
316
+ 0.7138387008487243,
317
+ 0.7231695330579867,
318
+ 0.7222552165311045
319
+ ],
320
+ "dead_neuron_fraction": 0.9765165088403593,
321
+ "l0_sparsity_base": 114.903941718445,
322
+ "l0_sparsity_aligned": 113.72704084477644,
323
+ "l0_base_by_layer": [
324
+ 105.1082713076474,
325
+ 113.19375729713953,
326
+ 126.40978911266784
327
+ ],
328
+ "l0_aligned_by_layer": [
329
+ 107.23830633391711,
330
+ 112.16951255107998,
331
+ 121.77329611792177
332
+ ],
333
+ "val_l0_base": 115.63809172145983,
334
+ "val_l0_aligned": 114.36136846891873,
335
+ "val_l0_base_by_layer": [
336
+ 105.9445899044656,
337
+ 113.89114311977207,
338
+ 127.07853403141361
339
+ ],
340
+ "val_l0_aligned_by_layer": [
341
+ 107.85962041884817,
342
+ 112.91301267933471,
343
+ 122.31146380040033
344
+ ],
345
+ "superposition_fraction": 0.0
346
+ }
v1/assembled_activations/llama32-3b-orpo/metrics/training_metrics.json ADDED
@@ -0,0 +1,524 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epochs": [
3
+ 1,
4
+ 2,
5
+ 3,
6
+ 4,
7
+ 5,
8
+ 6,
9
+ 7,
10
+ 8
11
+ ],
12
+ "train_loss": [
13
+ 0.705383880908769,
14
+ 0.5214350387079981,
15
+ 0.5085271640054956,
16
+ 0.5009160858980759,
17
+ 0.4938995702114985,
18
+ 0.4880766625370954,
19
+ 0.4838828326350266,
20
+ 0.4813817020438251
21
+ ],
22
+ "val_loss": [
23
+ 0.5289538633760982,
24
+ 0.5139043463774376,
25
+ 0.5094175137462417,
26
+ 0.505073328604873,
27
+ 0.5018799361134074,
28
+ 0.49963806452551435,
29
+ 0.49829376524031477,
30
+ 0.49799163226057724
31
+ ],
32
+ "train_fve_base": [
33
+ 0.49998613545751025,
34
+ 0.614935853087408,
35
+ 0.6300250220493622,
36
+ 0.6392686449500182,
37
+ 0.6469652943978585,
38
+ 0.6536180983839575,
39
+ 0.6583704736520865,
40
+ 0.6608418811501917
41
+ ],
42
+ "train_fve_aligned": [
43
+ 0.595036483206392,
44
+ 0.6966691696706461,
45
+ 0.70565621266919,
46
+ 0.7105856088957311,
47
+ 0.715512059233165,
48
+ 0.7198717664085629,
49
+ 0.7231761177102516,
50
+ 0.725106173172654
51
+ ],
52
+ "val_fve_base": [
53
+ 0.6004634630617671,
54
+ 0.619827045196014,
55
+ 0.6251299231464326,
56
+ 0.6303333112706688,
57
+ 0.634704968067988,
58
+ 0.63714382879397,
59
+ 0.6383371680818927,
60
+ 0.6388642060819096
61
+ ],
62
+ "val_fve_aligned": [
63
+ 0.6946163839070585,
64
+ 0.7050333519256552,
65
+ 0.7085146972646263,
66
+ 0.7120080927279607,
67
+ 0.7146814815036914,
68
+ 0.7174009016046974,
69
+ 0.719217080408366,
70
+ 0.7197545064681488
71
+ ],
72
+ "val_fve_base_by_layer": [
73
+ [
74
+ 0.5920717666286448,
75
+ 0.6040665744486904,
76
+ 0.6052519863188579
77
+ ],
78
+ [
79
+ 0.6119502233585138,
80
+ 0.6236264165783427,
81
+ 0.623904435422408
82
+ ],
83
+ [
84
+ 0.6160292325843691,
85
+ 0.6283739436983438,
86
+ 0.6309865357364035
87
+ ],
88
+ [
89
+ 0.6218419926952942,
90
+ 0.633468728727071,
91
+ 0.6356891599625193
92
+ ],
93
+ [
94
+ 0.6262971437414279,
95
+ 0.6373021262478454,
96
+ 0.6405155742979799
97
+ ],
98
+ [
99
+ 0.6280666652150179,
100
+ 0.6405123322421967,
101
+ 0.6428524321286466
102
+ ],
103
+ [
104
+ 0.6295256942354571,
105
+ 0.6411026166995782,
106
+ 0.6443831293370711
107
+ ],
108
+ [
109
+ 0.6299517154693604,
110
+ 0.6417424791146323,
111
+ 0.644898364681224
112
+ ]
113
+ ],
114
+ "val_fve_aligned_by_layer": [
115
+ [
116
+ 0.6938559163303275,
117
+ 0.6896499523317626,
118
+ 0.700343218149315
119
+ ],
120
+ [
121
+ 0.704806842416993,
122
+ 0.6989743828149366,
123
+ 0.7113187796782449
124
+ ],
125
+ [
126
+ 0.7070379946868457,
127
+ 0.7049488859026843,
128
+ 0.7135571484790423
129
+ ],
130
+ [
131
+ 0.7075730125317399,
132
+ 0.711481070331254,
133
+ 0.7169701475747593
134
+ ],
135
+ [
136
+ 0.7088280210944371,
137
+ 0.715514197399479,
138
+ 0.7197021714055726
139
+ ],
140
+ [
141
+ 0.711677071311711,
142
+ 0.7196013053674348,
143
+ 0.7209242694665
144
+ ],
145
+ [
146
+ 0.7131912720765119,
147
+ 0.722360813804946,
148
+ 0.7220990951148627
149
+ ],
150
+ [
151
+ 0.7138387008487243,
152
+ 0.7231695330579867,
153
+ 0.7222552165311045
154
+ ]
155
+ ],
156
+ "train_fve_base_by_layer": [
157
+ [
158
+ 0.4923095455381234,
159
+ 0.5029808437692026,
160
+ 0.5046679700457835
161
+ ],
162
+ [
163
+ 0.6069951350759917,
164
+ 0.6184060490068747,
165
+ 0.6194063158182811
166
+ ],
167
+ [
168
+ 0.6218264799889109,
169
+ 0.6333464354850782,
170
+ 0.634902091173839
171
+ ],
172
+ [
173
+ 0.6307406503765888,
174
+ 0.6425333459406517,
175
+ 0.6445318793805106
176
+ ],
177
+ [
178
+ 0.6380724335579725,
179
+ 0.6504222542934786,
180
+ 0.652401133614956
181
+ ],
182
+ [
183
+ 0.6442190956735361,
184
+ 0.6572884782948691,
185
+ 0.6593466577513205
186
+ ],
187
+ [
188
+ 0.6488628558644898,
189
+ 0.6620501859175554,
190
+ 0.6641983206830244
191
+ ],
192
+ [
193
+ 0.6513066935873282,
194
+ 0.6645377088692616,
195
+ 0.6666811821200453
196
+ ]
197
+ ],
198
+ "train_fve_aligned_by_layer": [
199
+ [
200
+ 0.5935017106411963,
201
+ 0.5907552899217577,
202
+ 0.6008523927400213
203
+ ],
204
+ [
205
+ 0.6946202010281658,
206
+ 0.6915539060499121,
207
+ 0.7038333473745592
208
+ ],
209
+ [
210
+ 0.7044359587885383,
211
+ 0.7012647485705196,
212
+ 0.711267870000003
213
+ ],
214
+ [
215
+ 0.7074584229536663,
216
+ 0.7087685695462803,
217
+ 0.7155297761831935
218
+ ],
219
+ [
220
+ 0.7092359761918976,
221
+ 0.716958020523982,
222
+ 0.7203421224924256
223
+ ],
224
+ [
225
+ 0.7133184386775011,
226
+ 0.7224183912104636,
227
+ 0.7238784134561947
228
+ ],
229
+ [
230
+ 0.7166503101013171,
231
+ 0.7264159107792816,
232
+ 0.7264620721235793
233
+ ],
234
+ [
235
+ 0.7181735973015767,
236
+ 0.7291909772543899,
237
+ 0.7279538861924421
238
+ ]
239
+ ],
240
+ "dead_neurons": [
241
+ 0.9320522252398489,
242
+ 0.9707254489750036,
243
+ 0.9742093351732133,
244
+ 0.9758863939712845,
245
+ 0.9764928740503074,
246
+ 0.9766615019927439,
247
+ 0.9765581015390885,
248
+ 0.9765165088403593
249
+ ],
250
+ "l0_base": [
251
+ 153.79902892995145,
252
+ 111.25570037806166,
253
+ 106.5397354856403,
254
+ 106.82006718356638,
255
+ 109.14992222560475,
256
+ 112.1244552576674,
257
+ 114.11547368768358,
258
+ 114.903941718445
259
+ ],
260
+ "l0_aligned": [
261
+ 135.68489490693867,
262
+ 102.3597272259411,
263
+ 101.5633843728693,
264
+ 103.30828229433183,
265
+ 106.6990623674543,
266
+ 110.25266599905679,
267
+ 112.71657555575129,
268
+ 113.72704084477644
269
+ ],
270
+ "l0_base_by_layer": [
271
+ [
272
+ 148.3754743140689,
273
+ 157.2706691476941,
274
+ 155.75093038528897
275
+ ],
276
+ [
277
+ 104.45322533566842,
278
+ 113.08530356100408,
279
+ 116.22856465265616
280
+ ],
281
+ [
282
+ 97.51032545242265,
283
+ 108.15526488616463,
284
+ 113.95360843549328
285
+ ],
286
+ [
287
+ 97.2745913601868,
288
+ 106.83384413309983,
289
+ 116.35175861062463
290
+ ],
291
+ [
292
+ 99.71026707530648,
293
+ 108.08216579100993,
294
+ 119.65732632807939
295
+ ],
296
+ [
297
+ 102.44186004086399,
298
+ 110.87372300058377,
299
+ 123.05777510215995
300
+ ],
301
+ [
302
+ 104.33490221833041,
303
+ 112.54172139521307,
304
+ 125.46978984238179
305
+ ],
306
+ [
307
+ 105.1082713076474,
308
+ 113.19375729713953,
309
+ 126.40978911266784
310
+ ]
311
+ ],
312
+ "l0_aligned_by_layer": [
313
+ [
314
+ 130.93926955633393,
315
+ 138.8549511091652,
316
+ 137.26045315236428
317
+ ],
318
+ [
319
+ 97.70284223584355,
320
+ 104.7684617629889,
321
+ 104.60786996497373
322
+ ],
323
+ [
324
+ 95.53064798598949,
325
+ 103.94935785172213,
326
+ 105.21013937536486
327
+ ],
328
+ [
329
+ 96.82353692352598,
330
+ 103.9993797431407,
331
+ 109.10192279626386
332
+ ],
333
+ [
334
+ 100.21243797431407,
335
+ 106.43509194395797,
336
+ 113.44964973730298
337
+ ],
338
+ [
339
+ 103.81118651488616,
340
+ 109.35150321074138,
341
+ 117.59530064214827
342
+ ],
343
+ [
344
+ 106.30086106246351,
345
+ 111.32382880910683,
346
+ 120.52502918855808
347
+ ],
348
+ [
349
+ 107.23830633391711,
350
+ 112.16951255107998,
351
+ 121.77329611792177
352
+ ]
353
+ ],
354
+ "val_l0_base": [
355
+ 116.11909428941017,
356
+ 109.00491089346521,
357
+ 106.40983377826151,
358
+ 108.02501710422376,
359
+ 111.56639294849016,
360
+ 113.66777837463698,
361
+ 114.9668982141305,
362
+ 115.63809172145983
363
+ ],
364
+ "val_l0_aligned": [
365
+ 103.58991680844292,
366
+ 101.88372834689955,
367
+ 101.8098484059279,
368
+ 105.87162105700108,
369
+ 108.83213614418868,
370
+ 112.11204097408275,
371
+ 113.55664873372822,
372
+ 114.36136846891873
373
+ ],
374
+ "val_l0_base_by_layer": [
375
+ [
376
+ 110.09255018783489,
377
+ 117.62194592540801,
378
+ 120.64277924542652
379
+ ],
380
+ [
381
+ 100.53130456415147,
382
+ 111.21847733902058,
383
+ 115.26494330760696
384
+ ],
385
+ [
386
+ 96.38699828142894,
387
+ 106.90630456415147,
388
+ 115.93619109947645
389
+ ],
390
+ [
391
+ 98.62003709882966,
392
+ 107.03457681546037,
393
+ 118.42042976898672
394
+ ],
395
+ [
396
+ 102.54924738219896,
397
+ 109.85809338155217,
398
+ 122.29183029254693
399
+ ],
400
+ [
401
+ 103.90275961191867,
402
+ 112.52176047120419,
403
+ 124.57880673233751
404
+ ],
405
+ [
406
+ 105.41001308900523,
407
+ 113.19371727748691,
408
+ 126.29695680628272
409
+ ],
410
+ [
411
+ 105.9445899044656,
412
+ 113.89114311977207,
413
+ 127.07853403141361
414
+ ]
415
+ ],
416
+ "val_l0_aligned_by_layer": [
417
+ [
418
+ 98.60367584228516,
419
+ 106.70844240837697,
420
+ 105.45762434554973
421
+ ],
422
+ [
423
+ 95.01434338155217,
424
+ 105.85416667998149,
425
+ 104.78266798888201
426
+ ],
427
+ [
428
+ 95.20391582069597,
429
+ 102.82864311977207,
430
+ 107.3969786479211
431
+ ],
432
+ [
433
+ 99.21869547579301,
434
+ 105.55650089923,
435
+ 112.83965968586388
436
+ ],
437
+ [
438
+ 102.3291339574684,
439
+ 108.66404887893437,
440
+ 115.50321772710191
441
+ ],
442
+ [
443
+ 105.50921686781639,
444
+ 111.07280759162303,
445
+ 119.75409031413612
446
+ ],
447
+ [
448
+ 106.91672120418848,
449
+ 112.04662958115183,
450
+ 121.70658814595008
451
+ ],
452
+ [
453
+ 107.85962041884817,
454
+ 112.91301267933471,
455
+ 122.31146380040033
456
+ ]
457
+ ],
458
+ "self_recon": [
459
+ 0.41878786202905055,
460
+ 0.3174672956033123,
461
+ 0.3066660691546336,
462
+ 0.30028960070481997,
463
+ 0.29444848812607666,
464
+ 0.2894557218278839,
465
+ 0.2858269191926081,
466
+ 0.2837859873213548
467
+ ],
468
+ "cross_recon": [
469
+ 0.47454479312381753,
470
+ 0.3836164794431676,
471
+ 0.37483930325062736,
472
+ 0.36926656394274837,
473
+ 0.3643547691800618,
474
+ 0.36018431452861394,
475
+ 0.35727764684348096,
476
+ 0.3555644095570319
477
+ ],
478
+ "sparsity": [
479
+ 0.09677809841117523,
480
+ 0.05052114697146054,
481
+ 0.05192536887888385,
482
+ 0.05291985494225469,
483
+ 0.053709169874762624,
484
+ 0.05454721025170044,
485
+ 0.05514484992978243,
486
+ 0.05536994681423405
487
+ ],
488
+ "val_self_recon": [
489
+ 0.323870322623178,
490
+ 0.31057615976059,
491
+ 0.30652437035325936,
492
+ 0.3025186664146903,
493
+ 0.2992763463115193,
494
+ 0.2968968311373476,
495
+ 0.2954498385883751,
496
+ 0.29497009999464946
497
+ ],
498
+ "val_cross_recon": [
499
+ 0.3886164392476307,
500
+ 0.3784953566433872,
501
+ 0.37661581997471955,
502
+ 0.37395940947283,
503
+ 0.37190580617694957,
504
+ 0.37063901352632733,
505
+ 0.36980168716445644,
506
+ 0.3697263550071816
507
+ ],
508
+ "val_sparsity": [
509
+ 0.049636956518857266,
510
+ 0.05193003809264817,
511
+ 0.052246809083753856,
512
+ 0.052970892627825915,
513
+ 0.05384126350439656,
514
+ 0.054485625367982224,
515
+ 0.05492324762397412,
516
+ 0.05513098687713683
517
+ ],
518
+ "layers": [
519
+ 24,
520
+ 25,
521
+ 26
522
+ ],
523
+ "topk_mode": "model_balanced_layer_agg"
524
+ }