ishangarg183 commited on
Commit
4fd97ae
·
verified ·
1 Parent(s): 374c918

Add assembled activations for llama32-3b-simpo

Browse files
Files changed (33) hide show
  1. .gitattributes +3 -0
  2. v1/assembled_activations/llama32-3b-simpo/activations/activations.pt +3 -0
  3. v1/assembled_activations/llama32-3b-simpo/checkpoints/final.pt +3 -0
  4. v1/assembled_activations/llama32-3b-simpo/features/counterfactual_scores.csv +0 -0
  5. v1/assembled_activations/llama32-3b-simpo/features/counterfactual_scores_by_layer.csv +0 -0
  6. v1/assembled_activations/llama32-3b-simpo/features/cross_layer_cosine_drift.csv +3 -0
  7. v1/assembled_activations/llama32-3b-simpo/features/decoder_layer_profiles.csv +3 -0
  8. v1/assembled_activations/llama32-3b-simpo/features/feature_activations.pt +3 -0
  9. v1/assembled_activations/llama32-3b-simpo/features/feature_classification.csv +0 -0
  10. v1/assembled_activations/llama32-3b-simpo/features/merged_classification.csv +0 -0
  11. v1/assembled_activations/llama32-3b-simpo/features/model_layer_stream_patterns.csv +3 -0
  12. v1/assembled_activations/llama32-3b-simpo/features/superposition_analysis.json +0 -0
  13. v1/assembled_activations/llama32-3b-simpo/metrics/aggregate_metrics.json +390 -0
  14. v1/assembled_activations/llama32-3b-simpo/metrics/training_metrics.json +524 -0
  15. v1/assembled_activations/llama32-3b-simpo/plots/aligned_decoder_norm_heatmap.png +3 -0
  16. v1/assembled_activations/llama32-3b-simpo/plots/base_decoder_norm_heatmap.png +3 -0
  17. v1/assembled_activations/llama32-3b-simpo/plots/cf_shift_by_layer.png +3 -0
  18. v1/assembled_activations/llama32-3b-simpo/plots/cf_shift_p95_by_layer.png +3 -0
  19. v1/assembled_activations/llama32-3b-simpo/plots/class_distribution_multilayer.png +3 -0
  20. v1/assembled_activations/llama32-3b-simpo/plots/class_distribution_primary.png +3 -0
  21. v1/assembled_activations/llama32-3b-simpo/plots/cross_layer_cosine_drift_by_stream.png +3 -0
  22. v1/assembled_activations/llama32-3b-simpo/plots/decoder_norm_ratio_by_layer.png +3 -0
  23. v1/assembled_activations/llama32-3b-simpo/plots/feature_layer_trajectories.png +3 -0
  24. v1/assembled_activations/llama32-3b-simpo/plots/feature_sharing_ratio_by_layer.png +3 -0
  25. v1/assembled_activations/llama32-3b-simpo/plots/fve_by_layer.png +3 -0
  26. v1/assembled_activations/llama32-3b-simpo/plots/l0_by_layer.png +3 -0
  27. v1/assembled_activations/llama32-3b-simpo/plots/layer_concentration_entropy.png +3 -0
  28. v1/assembled_activations/llama32-3b-simpo/plots/loss_curves.png +3 -0
  29. v1/assembled_activations/llama32-3b-simpo/plots/max_norm_layer_migration.png +3 -0
  30. v1/assembled_activations/llama32-3b-simpo/plots/rho_histogram_by_layer.png +3 -0
  31. v1/assembled_activations/llama32-3b-simpo/plots/rho_theta_scatter_by_layer.png +3 -0
  32. v1/assembled_activations/llama32-3b-simpo/plots/superposition_by_layer.png +3 -0
  33. v1/assembled_activations/llama32-3b-simpo/plots/theta_by_layer.png +3 -0
.gitattributes CHANGED
@@ -80,3 +80,6 @@ v1/assembled_activations/llama32-3b-kto/features/cross_layer_cosine_drift.csv fi
80
  v1/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
81
  v1/assembled_activations/llama32-3b-orpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
82
  v1/assembled_activations/llama32-3b-orpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
 
 
 
 
80
  v1/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
81
  v1/assembled_activations/llama32-3b-orpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
82
  v1/assembled_activations/llama32-3b-orpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
83
+ v1/assembled_activations/llama32-3b-simpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
84
+ v1/assembled_activations/llama32-3b-simpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
85
+ v1/assembled_activations/llama32-3b-simpo/features/model_layer_stream_patterns.csv filter=lfs diff=lfs merge=lfs -text
v1/assembled_activations/llama32-3b-simpo/activations/activations.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6a7e0d0b9cb1f470d02c3ca07896c10ffae2ecd2c427059f9e020bc764c17a1
3
+ size 4496133925
v1/assembled_activations/llama32-3b-simpo/checkpoints/final.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ccb6d15d31851a2c8616da9042c402d47e5863ff5af099931d9f1041b86f0d4
3
+ size 10873439922
v1/assembled_activations/llama32-3b-simpo/features/counterfactual_scores.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-simpo/features/counterfactual_scores_by_layer.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-simpo/features/cross_layer_cosine_drift.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fde65a9d5d2e756fabc40ffd1a57e4cc6bfe199673e000347a8b295cc77d0e3
3
+ size 27349788
v1/assembled_activations/llama32-3b-simpo/features/decoder_layer_profiles.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:419a4021e282db8cb1143e2576e64cb82eb8327da1cb2627676f9cdd475aa2a1
3
+ size 13236337
v1/assembled_activations/llama32-3b-simpo/features/feature_activations.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:544c4c9e0a04496f86034bd28fed57f5a7a8cd41185b3f7b4812d001f45f230a
3
+ size 35935153173
v1/assembled_activations/llama32-3b-simpo/features/feature_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-simpo/features/merged_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-simpo/features/model_layer_stream_patterns.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98183a1f8d1d6f341e6de2278123c273f14d27872b9ead0fd01199623ec143e5
3
+ size 10829188
v1/assembled_activations/llama32-3b-simpo/features/superposition_analysis.json ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-simpo/metrics/aggregate_metrics.json ADDED
@@ -0,0 +1,390 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crosscoder_kind": "multilayer_sparc",
3
+ "layers": [
4
+ 10,
5
+ 11,
6
+ 12
7
+ ],
8
+ "topk_mode": "model_balanced_layer_agg",
9
+ "topk": 400,
10
+ "expansion_factor": 8,
11
+ "dict_size": 24576,
12
+ "forced_shared_fraction": 0.06,
13
+ "class_counts": {
14
+ "aligned_only": 7032,
15
+ "shared_intermediate": 6309,
16
+ "base_only": 4631,
17
+ "shared_redirected": 3294,
18
+ "shared_attenuated": 1746,
19
+ "shared_aligned": 1564
20
+ },
21
+ "multilayer_class_counts": {
22
+ "drifting_or_rotating": 9101,
23
+ "persistent_aligned_only": 4366,
24
+ "persistent_shared": 3358,
25
+ "localized_aligned_only": 2666,
26
+ "persistent_base_only": 2377,
27
+ "localized_base_only": 2254,
28
+ "mixed_or_ambiguous": 454
29
+ },
30
+ "classification_thresholds": {
31
+ "rho_base_only": 0.4,
32
+ "rho_aligned_only": 0.5646282217167712,
33
+ "rho_shared_low": 0.43133134923386174,
34
+ "rho_shared_high": 0.5646282217167712
35
+ },
36
+ "threshold_sensitivity": {
37
+ "original": {
38
+ "aligned_only": 7032,
39
+ "shared_intermediate": 6309,
40
+ "base_only": 4631,
41
+ "shared_redirected": 3294,
42
+ "shared_attenuated": 1746,
43
+ "shared_aligned": 1564
44
+ },
45
+ "perturbed": {
46
+ "delta_-0.05": {
47
+ "base_only": 1938,
48
+ "aligned_only": 4114,
49
+ "shared_aligned": 1612,
50
+ "shared_redirected": 5669,
51
+ "shared_intermediate": 9535,
52
+ "shared_attenuated": 1708,
53
+ "other": 0
54
+ },
55
+ "delta_+0.05": {
56
+ "base_only": 7615,
57
+ "aligned_only": 10576,
58
+ "shared_aligned": 1486,
59
+ "shared_redirected": 872,
60
+ "shared_intermediate": 1690,
61
+ "shared_attenuated": 2337,
62
+ "other": 0
63
+ }
64
+ },
65
+ "perturbation": 0.05
66
+ },
67
+ "class_counts_by_layer": {
68
+ "10": {
69
+ "shared_intermediate": 7093,
70
+ "shared_redirected": 5890,
71
+ "base_only": 5623,
72
+ "shared_aligned": 3521,
73
+ "aligned_only": 2449
74
+ },
75
+ "11": {
76
+ "shared_attenuated": 8771,
77
+ "other": 5527,
78
+ "shared_intermediate": 4647,
79
+ "shared_redirected": 2565,
80
+ "shared_aligned": 1389,
81
+ "base_only": 1060,
82
+ "aligned_only": 617
83
+ },
84
+ "12": {
85
+ "shared_redirected": 9538,
86
+ "shared_intermediate": 5957,
87
+ "base_only": 3856,
88
+ "shared_aligned": 3719,
89
+ "aligned_only": 1506
90
+ }
91
+ },
92
+ "feature_sharing_ratio_by_layer": {
93
+ "10": 0.6715494791666666,
94
+ "11": 0.7068684895833334,
95
+ "12": 0.7818196614583334
96
+ },
97
+ "decoder_amplification_by_layer": {
98
+ "10": {
99
+ "median": 0.9999999870874021,
100
+ "p95": 5.846633247470493
101
+ },
102
+ "11": {
103
+ "median": 0.9999999890566127,
104
+ "p95": 2.8997828111006294
105
+ },
106
+ "12": {
107
+ "median": 0.999999988812232,
108
+ "p95": 3.348760753383095
109
+ }
110
+ },
111
+ "classification_thresholds_by_layer": {
112
+ "10": {
113
+ "rho_base_only": 0.3298648264878633,
114
+ "rho_aligned_only": 0.7696755668389346,
115
+ "rho_shared_low": 0.3298648264878633,
116
+ "rho_shared_high": 0.7696755668389346
117
+ },
118
+ "11": {
119
+ "rho_base_only": 0.2341652512550354,
120
+ "rho_aligned_only": 0.8046609088778496,
121
+ "rho_shared_low": 0.46459993720054626,
122
+ "rho_shared_high": 0.6093218177556992
123
+ },
124
+ "12": {
125
+ "rho_base_only": 0.33508244103501,
126
+ "rho_aligned_only": 0.7473176242691034,
127
+ "rho_shared_low": 0.33508244103501,
128
+ "rho_shared_high": 0.7473176242691034
129
+ }
130
+ },
131
+ "threshold_sensitivity_by_layer": {
132
+ "10": {
133
+ "original": {
134
+ "shared_intermediate": 7093,
135
+ "shared_redirected": 5890,
136
+ "base_only": 5623,
137
+ "shared_aligned": 3521,
138
+ "aligned_only": 2449
139
+ },
140
+ "perturbed": {
141
+ "delta_-0.05": {
142
+ "base_only": 3901,
143
+ "aligned_only": 1675,
144
+ "shared_aligned": 4473,
145
+ "shared_redirected": 6604,
146
+ "shared_intermediate": 7923,
147
+ "shared_attenuated": 0,
148
+ "other": 0
149
+ },
150
+ "delta_+0.05": {
151
+ "base_only": 7427,
152
+ "aligned_only": 3600,
153
+ "shared_aligned": 2604,
154
+ "shared_redirected": 4977,
155
+ "shared_intermediate": 5968,
156
+ "shared_attenuated": 0,
157
+ "other": 0
158
+ }
159
+ },
160
+ "perturbation": 0.05
161
+ },
162
+ "11": {
163
+ "original": {
164
+ "shared_attenuated": 8771,
165
+ "other": 5527,
166
+ "shared_intermediate": 4647,
167
+ "shared_redirected": 2565,
168
+ "shared_aligned": 1389,
169
+ "base_only": 1060,
170
+ "aligned_only": 617
171
+ },
172
+ "perturbed": {
173
+ "delta_-0.05": {
174
+ "base_only": 637,
175
+ "aligned_only": 374,
176
+ "shared_aligned": 2466,
177
+ "shared_redirected": 3929,
178
+ "shared_intermediate": 7325,
179
+ "shared_attenuated": 6633,
180
+ "other": 3212
181
+ },
182
+ "delta_+0.05": {
183
+ "base_only": 1818,
184
+ "aligned_only": 1085,
185
+ "shared_aligned": 54,
186
+ "shared_redirected": 814,
187
+ "shared_intermediate": 1509,
188
+ "shared_attenuated": 11833,
189
+ "other": 7463
190
+ }
191
+ },
192
+ "perturbation": 0.05
193
+ },
194
+ "12": {
195
+ "original": {
196
+ "shared_redirected": 9538,
197
+ "shared_intermediate": 5957,
198
+ "base_only": 3856,
199
+ "shared_aligned": 3719,
200
+ "aligned_only": 1506
201
+ },
202
+ "perturbed": {
203
+ "delta_-0.05": {
204
+ "base_only": 2386,
205
+ "aligned_only": 972,
206
+ "shared_aligned": 4290,
207
+ "shared_redirected": 10373,
208
+ "shared_intermediate": 6555,
209
+ "shared_attenuated": 0,
210
+ "other": 0
211
+ },
212
+ "delta_+0.05": {
213
+ "base_only": 6109,
214
+ "aligned_only": 2494,
215
+ "shared_aligned": 2781,
216
+ "shared_redirected": 8206,
217
+ "shared_intermediate": 4986,
218
+ "shared_attenuated": 0,
219
+ "other": 0
220
+ }
221
+ },
222
+ "perturbation": 0.05
223
+ }
224
+ },
225
+ "counterfactual_shift_by_layer": {
226
+ "10": {
227
+ "aligned_only": {
228
+ "mean_shift": 3.2320919899215418e-06,
229
+ "median_shift": 0.0,
230
+ "p95_abs_shift": 0.0,
231
+ "count": 7032
232
+ },
233
+ "base_only": {
234
+ "mean_shift": -3.5814365136129003e-06,
235
+ "median_shift": 0.0,
236
+ "p95_abs_shift": 0.0,
237
+ "count": 4631
238
+ },
239
+ "shared_aligned": {
240
+ "mean_shift": 5.724467744014439e-05,
241
+ "median_shift": 0.0,
242
+ "p95_abs_shift": 0.07152953743934631,
243
+ "count": 1564
244
+ },
245
+ "shared_attenuated": {
246
+ "mean_shift": -4.0716883922622904e-07,
247
+ "median_shift": 0.0,
248
+ "p95_abs_shift": 0.0,
249
+ "count": 1746
250
+ },
251
+ "shared_intermediate": {
252
+ "mean_shift": -7.30478225800923e-08,
253
+ "median_shift": 0.0,
254
+ "p95_abs_shift": 0.0,
255
+ "count": 6309
256
+ },
257
+ "shared_redirected": {
258
+ "mean_shift": 4.9160252745138344e-08,
259
+ "median_shift": 0.0,
260
+ "p95_abs_shift": 0.0,
261
+ "count": 3294
262
+ }
263
+ },
264
+ "11": {
265
+ "aligned_only": {
266
+ "mean_shift": 1.867720686331114e-05,
267
+ "median_shift": 0.0,
268
+ "p95_abs_shift": 0.0,
269
+ "count": 7032
270
+ },
271
+ "base_only": {
272
+ "mean_shift": -1.9894835705085598e-05,
273
+ "median_shift": 0.0,
274
+ "p95_abs_shift": 0.0,
275
+ "count": 4631
276
+ },
277
+ "shared_aligned": {
278
+ "mean_shift": 7.235741871226919e-05,
279
+ "median_shift": 0.0,
280
+ "p95_abs_shift": 0.0822630167007446,
281
+ "count": 1564
282
+ },
283
+ "shared_attenuated": {
284
+ "mean_shift": -3.471290511984925e-06,
285
+ "median_shift": 0.0,
286
+ "p95_abs_shift": 0.0,
287
+ "count": 1746
288
+ },
289
+ "shared_intermediate": {
290
+ "mean_shift": 8.675455243721306e-08,
291
+ "median_shift": 0.0,
292
+ "p95_abs_shift": 0.0,
293
+ "count": 6309
294
+ },
295
+ "shared_redirected": {
296
+ "mean_shift": -3.6504596481600716e-07,
297
+ "median_shift": 0.0,
298
+ "p95_abs_shift": 0.0,
299
+ "count": 3294
300
+ }
301
+ },
302
+ "12": {
303
+ "aligned_only": {
304
+ "mean_shift": 1.7178416289861293e-05,
305
+ "median_shift": 0.0,
306
+ "p95_abs_shift": 0.0,
307
+ "count": 7032
308
+ },
309
+ "base_only": {
310
+ "mean_shift": -1.7390496993661782e-05,
311
+ "median_shift": 0.0,
312
+ "p95_abs_shift": 0.0,
313
+ "count": 4631
314
+ },
315
+ "shared_aligned": {
316
+ "mean_shift": 0.00016059010712473373,
317
+ "median_shift": 0.0,
318
+ "p95_abs_shift": 0.08721660748124122,
319
+ "count": 1564
320
+ },
321
+ "shared_attenuated": {
322
+ "mean_shift": -4.775544404814352e-06,
323
+ "median_shift": 0.0,
324
+ "p95_abs_shift": 0.0,
325
+ "count": 1746
326
+ },
327
+ "shared_intermediate": {
328
+ "mean_shift": -4.4374906589709826e-08,
329
+ "median_shift": 0.0,
330
+ "p95_abs_shift": 0.0,
331
+ "count": 6309
332
+ },
333
+ "shared_redirected": {
334
+ "mean_shift": 3.799928669689255e-07,
335
+ "median_shift": 0.0,
336
+ "p95_abs_shift": 0.0,
337
+ "count": 3294
338
+ }
339
+ }
340
+ },
341
+ "total_features": 24576,
342
+ "fve_base": 0.6715279371326507,
343
+ "fve_aligned": 0.662245636523082,
344
+ "fve_base_by_layer": [
345
+ 0.6305625454293495,
346
+ 0.6942680309580259,
347
+ 0.6897531785265938
348
+ ],
349
+ "fve_aligned_by_layer": [
350
+ 0.6188769243774613,
351
+ 0.6861868694190579,
352
+ 0.6816730599128763
353
+ ],
354
+ "val_fve_base_by_layer": [
355
+ 0.6305625454293495,
356
+ 0.6942680309580259,
357
+ 0.6897531785265938
358
+ ],
359
+ "val_fve_aligned_by_layer": [
360
+ 0.6188769243774613,
361
+ 0.6861868694190579,
362
+ 0.6816730599128763
363
+ ],
364
+ "dead_neuron_fraction": 0.9593874006958921,
365
+ "l0_sparsity_base": 64.82328354345033,
366
+ "l0_sparsity_aligned": 63.5290322557083,
367
+ "l0_base_by_layer": [
368
+ 47.34114127262113,
369
+ 68.14484821949796,
370
+ 78.9838550788091
371
+ ],
372
+ "l0_aligned_by_layer": [
373
+ 46.07039915353182,
374
+ 66.73431115002919,
375
+ 77.78238105662581
376
+ ],
377
+ "val_l0_base": 66.31130234483649,
378
+ "val_l0_aligned": 64.88672773750665,
379
+ "val_l0_base_by_layer": [
380
+ 48.44813481675393,
381
+ 69.85569371727749,
382
+ 80.6300719895288
383
+ ],
384
+ "val_l0_aligned_by_layer": [
385
+ 46.990892240514306,
386
+ 68.39632419765931,
387
+ 79.27296032331377
388
+ ],
389
+ "superposition_fraction": 0.0
390
+ }
v1/assembled_activations/llama32-3b-simpo/metrics/training_metrics.json ADDED
@@ -0,0 +1,524 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epochs": [
3
+ 1,
4
+ 2,
5
+ 3,
6
+ 4,
7
+ 5,
8
+ 6,
9
+ 7,
10
+ 8
11
+ ],
12
+ "train_loss": [
13
+ 0.05927932382030092,
14
+ 0.03255366562918412,
15
+ 0.03115293783746264,
16
+ 0.030371403902273254,
17
+ 0.02979045822211719,
18
+ 0.02931892105268931,
19
+ 0.028975081874163124,
20
+ 0.02878278101344799
21
+ ],
22
+ "val_loss": [
23
+ 0.033846099475759485,
24
+ 0.032007066728010854,
25
+ 0.031248764461637792,
26
+ 0.030814522370227967,
27
+ 0.030503301778856996,
28
+ 0.030317795097203778,
29
+ 0.030241612618038167,
30
+ 0.030229815228291206
31
+ ],
32
+ "train_fve_base": [
33
+ 0.503612844223043,
34
+ 0.6349237019808336,
35
+ 0.6552128129303282,
36
+ 0.6672707846550238,
37
+ 0.6764131289647465,
38
+ 0.6837705131587548,
39
+ 0.688966876536312,
40
+ 0.6917263760388559
41
+ ],
42
+ "train_fve_aligned": [
43
+ 0.49909401890992744,
44
+ 0.6284301312639222,
45
+ 0.6463590448736379,
46
+ 0.6569368737661135,
47
+ 0.6652233918299121,
48
+ 0.6720354633935348,
49
+ 0.676885349633606,
50
+ 0.6794380337191268
51
+ ],
52
+ "val_fve_base": [
53
+ 0.6182141828287334,
54
+ 0.6435623983438102,
55
+ 0.6547190727987838,
56
+ 0.6613033850779708,
57
+ 0.6664242657067264,
58
+ 0.6696532880448546,
59
+ 0.671212948429647,
60
+ 0.6715279371326507
61
+ ],
62
+ "val_fve_aligned": [
63
+ 0.613135617752974,
64
+ 0.6363836629852575,
65
+ 0.6465314929397943,
66
+ 0.6526829559765561,
67
+ 0.6575311690724958,
68
+ 0.6605612236167748,
69
+ 0.6619636584950991,
70
+ 0.662245636523082
71
+ ],
72
+ "val_fve_base_by_layer": [
73
+ [
74
+ 0.5780349083595875,
75
+ 0.6430585827502905,
76
+ 0.633548992466552
77
+ ],
78
+ [
79
+ 0.6026088904335861,
80
+ 0.6670344584275291,
81
+ 0.6610437787640158
82
+ ],
83
+ [
84
+ 0.6138415374056831,
85
+ 0.6778382768181606,
86
+ 0.6724773370782743
87
+ ],
88
+ [
89
+ 0.6206396946108154,
90
+ 0.6839593848632892,
91
+ 0.6793110280136787
92
+ ],
93
+ [
94
+ 0.6254869090949053,
95
+ 0.6891793920731669,
96
+ 0.6846064441491172
97
+ ],
98
+ [
99
+ 0.6288681657526506,
100
+ 0.6923003037562545,
101
+ 0.6877913343968816
102
+ ],
103
+ [
104
+ 0.6302341993566583,
105
+ 0.6938901640357772,
106
+ 0.6895144166746688
107
+ ],
108
+ [
109
+ 0.6305625454293495,
110
+ 0.6942680309580259,
111
+ 0.6897531785265938
112
+ ]
113
+ ],
114
+ "val_fve_aligned_by_layer": [
115
+ [
116
+ 0.5722053659523969,
117
+ 0.6384917887717642,
118
+ 0.6287096414266456
119
+ ],
120
+ [
121
+ 0.5942014565018459,
122
+ 0.6606073991166359,
123
+ 0.6543420678033879
124
+ ],
125
+ [
126
+ 0.6036381100484838,
127
+ 0.6709235947169558,
128
+ 0.6650327238112844
129
+ ],
130
+ [
131
+ 0.609793980708297,
132
+ 0.676685466816288,
133
+ 0.6715693632969681
134
+ ],
135
+ [
136
+ 0.6141581582149286,
137
+ 0.6816896346851169,
138
+ 0.6767456603299885
139
+ ],
140
+ [
141
+ 0.6174167641170362,
142
+ 0.6844394194518084,
143
+ 0.6798274301733646
144
+ ],
145
+ [
146
+ 0.6185780351698711,
147
+ 0.6858815141997412,
148
+ 0.6814313761850926
149
+ ],
150
+ [
151
+ 0.6188769243774613,
152
+ 0.6861868694190579,
153
+ 0.6816730599128763
154
+ ]
155
+ ],
156
+ "train_fve_base_by_layer": [
157
+ [
158
+ 0.4677472282277167,
159
+ 0.5262350366258928,
160
+ 0.5168562211700858
161
+ ],
162
+ [
163
+ 0.5941825219920972,
164
+ 0.6585408136709628,
165
+ 0.6520477089350222
166
+ ],
167
+ [
168
+ 0.6142003669491026,
169
+ 0.6779960200094857,
170
+ 0.6734419917406148
171
+ ],
172
+ [
173
+ 0.6260286404664797,
174
+ 0.6899296233894807,
175
+ 0.6858540300869204
176
+ ],
177
+ [
178
+ 0.6346775183744481,
179
+ 0.6991928598369971,
180
+ 0.6953689465032846
181
+ ],
182
+ [
183
+ 0.6413471689714163,
184
+ 0.7068967281623814,
185
+ 0.7030675794670475
186
+ ],
187
+ [
188
+ 0.645974550411427,
189
+ 0.7123980270932448,
190
+ 0.7085279913861642
191
+ ],
192
+ [
193
+ 0.6484792617699311,
194
+ 0.7152978126514188,
195
+ 0.7114019931510952
196
+ ]
197
+ ],
198
+ "train_fve_aligned_by_layer": [
199
+ [
200
+ 0.4632670403361947,
201
+ 0.5220091638687544,
202
+ 0.5120058040554872
203
+ ],
204
+ [
205
+ 0.5868167219214876,
206
+ 0.6527098008087345,
207
+ 0.6457638119440363
208
+ ],
209
+ [
210
+ 0.6038263409164171,
211
+ 0.6701786429716423,
212
+ 0.6650720916153462
213
+ ],
214
+ [
215
+ 0.6137546507921124,
216
+ 0.6809061225297085,
217
+ 0.6761497881978976
218
+ ],
219
+ [
220
+ 0.6210833272249185,
221
+ 0.6895746217091322,
222
+ 0.685012164898108
223
+ ],
224
+ [
225
+ 0.6268049689043493,
226
+ 0.6969297279479276,
227
+ 0.692371636472525
228
+ ],
229
+ [
230
+ 0.6309437568745833,
231
+ 0.7021611605689878,
232
+ 0.6975510722353524
233
+ ],
234
+ [
235
+ 0.6331689965634477,
236
+ 0.7048862822836627,
237
+ 0.7002587634363302
238
+ ]
239
+ ],
240
+ "dead_neurons": [
241
+ 0.8715425254097475,
242
+ 0.9557898779118611,
243
+ 0.961711463702748,
244
+ 0.9616781369817236,
245
+ 0.9606253722224817,
246
+ 0.9598153939876234,
247
+ 0.9594147409254253,
248
+ 0.9593874006958921
249
+ ],
250
+ "l0_base": [
251
+ 110.7568565675966,
252
+ 59.56439248623378,
253
+ 59.25963353136444,
254
+ 60.706109090917515,
255
+ 62.3549953445105,
256
+ 63.71940258268856,
257
+ 64.55443870290566,
258
+ 64.82328354345033
259
+ ],
260
+ "l0_aligned": [
261
+ 109.46531757414863,
262
+ 58.278739833442074,
263
+ 58.12394318856207,
264
+ 59.642915827032425,
265
+ 61.2593235343225,
266
+ 62.529463801965754,
267
+ 63.29711094114144,
268
+ 63.5290322557083
269
+ ],
270
+ "l0_base_by_layer": [
271
+ [
272
+ 93.83470154699359,
273
+ 118.0552393461763,
274
+ 120.38061879743141
275
+ ],
276
+ [
277
+ 43.45864346176299,
278
+ 63.35460449503795,
279
+ 71.87992556917688
280
+ ],
281
+ [
282
+ 43.179728546409805,
283
+ 62.33630691768827,
284
+ 72.2628612084063
285
+ ],
286
+ [
287
+ 44.387076765907764,
288
+ 63.711981903093985,
289
+ 74.01926444833626
290
+ ],
291
+ [
292
+ 45.60799766491535,
293
+ 65.49543928779919,
294
+ 75.96154407472271
295
+ ],
296
+ [
297
+ 46.5932209573847,
298
+ 66.9508720081728,
299
+ 77.61410901926445
300
+ ],
301
+ [
302
+ 47.15878575598366,
303
+ 67.86980078809107,
304
+ 78.63472343841214
305
+ ],
306
+ [
307
+ 47.34114127262113,
308
+ 68.14484821949796,
309
+ 78.9838550788091
310
+ ]
311
+ ],
312
+ "l0_aligned_by_layer": [
313
+ [
314
+ 92.87680604203152,
315
+ 116.69826328079392,
316
+ 118.8208734676007
317
+ ],
318
+ [
319
+ 42.17456582019848,
320
+ 62.11866973146527,
321
+ 70.5429801517805
322
+ ],
323
+ [
324
+ 41.8616097489784,
325
+ 61.297577349678924,
326
+ 71.2126386456509
327
+ ],
328
+ [
329
+ 43.14568739054291,
330
+ 62.718056771745474,
331
+ 73.06499927028605
332
+ ],
333
+ [
334
+ 44.36987375948628,
335
+ 64.42671847635727,
336
+ 74.98137405137186
337
+ ],
338
+ [
339
+ 45.319158639813196,
340
+ 65.73668272037361,
341
+ 76.53254524226503
342
+ ],
343
+ [
344
+ 45.89940893169877,
345
+ 66.51408347927612,
346
+ 77.47783493870404
347
+ ],
348
+ [
349
+ 46.07039915353182,
350
+ 66.73431115002919,
351
+ 77.78238105662581
352
+ ]
353
+ ],
354
+ "val_l0_base": [
355
+ 62.84433352884822,
356
+ 60.29535830832277,
357
+ 61.20708046159195,
358
+ 62.52763392663127,
359
+ 64.26254554069479,
360
+ 65.39581746206234,
361
+ 66.14812595307515,
362
+ 66.31130234483649
363
+ ],
364
+ "val_l0_aligned": [
365
+ 61.65801482674963,
366
+ 58.96664246963581,
367
+ 59.968006054144254,
368
+ 61.42377619718382,
369
+ 63.13096464866119,
370
+ 64.15648820887061,
371
+ 64.74734806639987,
372
+ 64.88672773750665
373
+ ],
374
+ "val_l0_base_by_layer": [
375
+ [
376
+ 46.69628055432704,
377
+ 67.60760254385583,
378
+ 74.2291121557745
379
+ ],
380
+ [
381
+ 44.08933246073298,
382
+ 63.24301920386509,
383
+ 73.55371948561744
384
+ ],
385
+ [
386
+ 44.62358203109022,
387
+ 64.35722079451796,
388
+ 74.6404341453033
389
+ ],
390
+ [
391
+ 45.78910340314136,
392
+ 65.5068172035417,
393
+ 76.28697643979058
394
+ ],
395
+ [
396
+ 46.838787098829656,
397
+ 67.65003272251309,
398
+ 78.29881110865408
399
+ ],
400
+ [
401
+ 47.84876746532181,
402
+ 68.77879581151832,
403
+ 79.55988219895288
404
+ ],
405
+ [
406
+ 48.298047570033845,
407
+ 69.62587262697869,
408
+ 80.52045157068063
409
+ ],
410
+ [
411
+ 48.44813481675393,
412
+ 69.85569371727749,
413
+ 80.6300719895288
414
+ ]
415
+ ],
416
+ "val_l0_aligned_by_layer": [
417
+ [
418
+ 45.42937391590698,
419
+ 66.43193717277487,
420
+ 73.11272905759162
421
+ ],
422
+ [
423
+ 42.52116056512163,
424
+ 62.154504805959334,
425
+ 72.22425830301815
426
+ ],
427
+ [
428
+ 43.128981245749905,
429
+ 63.41393979057592,
430
+ 73.36109293193718
431
+ ],
432
+ [
433
+ 44.41513962271326,
434
+ 64.71924083769633,
435
+ 75.13694371727749
436
+ ],
437
+ [
438
+ 45.48292976898673,
439
+ 66.78866712959649,
440
+ 77.12129147514622
441
+ ],
442
+ [
443
+ 46.53888526636893,
444
+ 67.60274869109948,
445
+ 78.32782505694485
446
+ ],
447
+ [
448
+ 46.87685428239912,
449
+ 68.23958335996298,
450
+ 79.12559992605479
451
+ ],
452
+ [
453
+ 46.990892240514306,
454
+ 68.39632419765931,
455
+ 79.27296032331377
456
+ ]
457
+ ],
458
+ "self_recon": [
459
+ 0.027993987802083356,
460
+ 0.020580280668407207,
461
+ 0.019489796229263563,
462
+ 0.018846164809260463,
463
+ 0.018348425166802105,
464
+ 0.017941102601361603,
465
+ 0.017650670092498533,
466
+ 0.01749732045938467
467
+ ],
468
+ "cross_recon": [
469
+ 0.028028677605305648,
470
+ 0.02058453171394909,
471
+ 0.019495135330868548,
472
+ 0.018860481901260683,
473
+ 0.018386862897496265,
474
+ 0.018007030048376626,
475
+ 0.01773270223207089,
476
+ 0.017584875620731955
477
+ ],
478
+ "sparsity": [
479
+ 0.020073864794261775,
480
+ 0.0037395720159709976,
481
+ 0.00386508731269351,
482
+ 0.0039810461591020165,
483
+ 0.004087287744385409,
484
+ 0.004175006297253338,
485
+ 0.004231330718039396,
486
+ 0.004251510163788699
487
+ ],
488
+ "val_self_recon": [
489
+ 0.021518522912529127,
490
+ 0.020130771841288236,
491
+ 0.01952393066274558,
492
+ 0.019163148765359562,
493
+ 0.01887795397358415,
494
+ 0.018701545738112862,
495
+ 0.018615932642672386,
496
+ 0.018599209112373633
497
+ ],
498
+ "val_cross_recon": [
499
+ 0.021522300454412457,
500
+ 0.020133654884408905,
501
+ 0.019527446098274587,
502
+ 0.019169504778625453,
503
+ 0.018888392441793887,
504
+ 0.01871574230745669,
505
+ 0.01863224525979832,
506
+ 0.018615821108261016
507
+ ],
508
+ "val_sparsity": [
509
+ 0.0037186559794363397,
510
+ 0.003822832755959003,
511
+ 0.003913855225247629,
512
+ 0.003983571632955399,
513
+ 0.004069990641803174,
514
+ 0.004129952197670156,
515
+ 0.004172781662752179,
516
+ 0.004184277574604874
517
+ ],
518
+ "layers": [
519
+ 10,
520
+ 11,
521
+ 12
522
+ ],
523
+ "topk_mode": "model_balanced_layer_agg"
524
+ }
v1/assembled_activations/llama32-3b-simpo/plots/aligned_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: 6d148f8c3598a04c1a3c8a3614f9ac7d755dbc992e11380fde57b5b27a6160f4
  • Pointer size: 131 Bytes
  • Size of remote file: 181 kB
v1/assembled_activations/llama32-3b-simpo/plots/base_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: 90378eba9ab620720e9a5be63ddf19d857f163754dae2a9c0e11761109eef681
  • Pointer size: 131 Bytes
  • Size of remote file: 180 kB
v1/assembled_activations/llama32-3b-simpo/plots/cf_shift_by_layer.png ADDED

Git LFS Details

  • SHA256: d2b279e1177392223a89e84117484da01bcfca14b6e209d50dfe53e7b52ce725
  • Pointer size: 130 Bytes
  • Size of remote file: 43.7 kB
v1/assembled_activations/llama32-3b-simpo/plots/cf_shift_p95_by_layer.png ADDED

Git LFS Details

  • SHA256: 851ff1df0a1a1053cce768efd353a568e5f95567794ab2f62307710fbf3079b4
  • Pointer size: 131 Bytes
  • Size of remote file: 125 kB
v1/assembled_activations/llama32-3b-simpo/plots/class_distribution_multilayer.png ADDED

Git LFS Details

  • SHA256: ad3199c1872d52c1b76cad0bfbecc17df5b3e7214c28e517a1c98dc434feb566
  • Pointer size: 130 Bytes
  • Size of remote file: 87.1 kB
v1/assembled_activations/llama32-3b-simpo/plots/class_distribution_primary.png ADDED

Git LFS Details

  • SHA256: 339f0f6aba2c93aae7ec7a30bc58f71b369d7f8067bf58a8f2d3e737fec8638b
  • Pointer size: 131 Bytes
  • Size of remote file: 148 kB
v1/assembled_activations/llama32-3b-simpo/plots/cross_layer_cosine_drift_by_stream.png ADDED

Git LFS Details

  • SHA256: 533c9c7ee417e3884b79be016c7d37e19499171e41899c787d94f7087e672374
  • Pointer size: 130 Bytes
  • Size of remote file: 50.2 kB
v1/assembled_activations/llama32-3b-simpo/plots/decoder_norm_ratio_by_layer.png ADDED

Git LFS Details

  • SHA256: 644a033382361decd1c16c9c403e22528a058b38708031dcf936d0fc13c72d9a
  • Pointer size: 130 Bytes
  • Size of remote file: 51.4 kB
v1/assembled_activations/llama32-3b-simpo/plots/feature_layer_trajectories.png ADDED

Git LFS Details

  • SHA256: 73106568eb9aeee5a10abf229d4bed081a19f449bdd8ae642ee8b248f0f14f34
  • Pointer size: 131 Bytes
  • Size of remote file: 658 kB
v1/assembled_activations/llama32-3b-simpo/plots/feature_sharing_ratio_by_layer.png ADDED

Git LFS Details

  • SHA256: 546e91aeafe39f3a08115604051f0e363f6a1d19c9c18cac97fd3ab2e2ae0e7c
  • Pointer size: 130 Bytes
  • Size of remote file: 55.3 kB
v1/assembled_activations/llama32-3b-simpo/plots/fve_by_layer.png ADDED

Git LFS Details

  • SHA256: 29a2c6dc652c68c5db0866990841863f7fbb73177badb1253971d775b81d05a2
  • Pointer size: 131 Bytes
  • Size of remote file: 111 kB
v1/assembled_activations/llama32-3b-simpo/plots/l0_by_layer.png ADDED

Git LFS Details

  • SHA256: 2fa4c740150bb2c6e395c42a3fbc42e81add40686df232319a3821d4afb4008f
  • Pointer size: 130 Bytes
  • Size of remote file: 93.7 kB
v1/assembled_activations/llama32-3b-simpo/plots/layer_concentration_entropy.png ADDED

Git LFS Details

  • SHA256: 77c88abe809a20c6260529ec09c812d3fc7f327ceca3fa0e4d339e2339e3447d
  • Pointer size: 131 Bytes
  • Size of remote file: 106 kB
v1/assembled_activations/llama32-3b-simpo/plots/loss_curves.png ADDED

Git LFS Details

  • SHA256: 7da65a8bfbcc7548869c6d28d685ec0edbe575c6a0080e4df97fc2599401aeb1
  • Pointer size: 131 Bytes
  • Size of remote file: 301 kB
v1/assembled_activations/llama32-3b-simpo/plots/max_norm_layer_migration.png ADDED

Git LFS Details

  • SHA256: 90905d06c7c2260d8581e90c8c7361d50b4c1d04b09b391d04d1f6f4e790c24f
  • Pointer size: 130 Bytes
  • Size of remote file: 83 kB
v1/assembled_activations/llama32-3b-simpo/plots/rho_histogram_by_layer.png ADDED

Git LFS Details

  • SHA256: db116820ff11ebc613106d1ed76dc035f2361577131214fbe84d8b0c208dfb91
  • Pointer size: 130 Bytes
  • Size of remote file: 48.5 kB
v1/assembled_activations/llama32-3b-simpo/plots/rho_theta_scatter_by_layer.png ADDED

Git LFS Details

  • SHA256: 1ffbb671d608deaf20429fb7b7161cfe947d4dc7099bf184651c4109022125d3
  • Pointer size: 132 Bytes
  • Size of remote file: 1.44 MB
v1/assembled_activations/llama32-3b-simpo/plots/superposition_by_layer.png ADDED

Git LFS Details

  • SHA256: f2c091fab892c52dd76137c531d910fceb852741343c66a11f6d205b1374da37
  • Pointer size: 131 Bytes
  • Size of remote file: 104 kB
v1/assembled_activations/llama32-3b-simpo/plots/theta_by_layer.png ADDED

Git LFS Details

  • SHA256: 86bb23a3595d6f9caa4c9b260c2954a09e6f89e81dd88466fa6fb9397201b486
  • Pointer size: 130 Bytes
  • Size of remote file: 53.2 kB