ishangarg183 commited on
Commit
a5b7df1
·
verified ·
1 Parent(s): c96aa0e

Rename v2-samelayer -> v1-undetermined (2/14)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +9 -0
  2. v1-undetermined/assembled_activations/llama32-3b-simpo/checkpoints/final.pt +3 -0
  3. v1-undetermined/assembled_activations/llama32-3b-simpo/features/counterfactual_scores.csv +0 -0
  4. v1-undetermined/assembled_activations/llama32-3b-simpo/features/counterfactual_scores_by_layer.csv +0 -0
  5. v1-undetermined/assembled_activations/llama32-3b-simpo/features/cross_layer_cosine_drift.csv +3 -0
  6. v1-undetermined/assembled_activations/llama32-3b-simpo/features/decoder_layer_profiles.csv +3 -0
  7. v1-undetermined/assembled_activations/llama32-3b-simpo/features/feature_classification.csv +0 -0
  8. v1-undetermined/assembled_activations/llama32-3b-simpo/features/merged_classification.csv +0 -0
  9. v1-undetermined/assembled_activations/llama32-3b-simpo/features/model_layer_stream_patterns.csv +3 -0
  10. v1-undetermined/assembled_activations/llama32-3b-simpo/features/superposition_analysis.json +0 -0
  11. v1-undetermined/assembled_activations/llama32-3b-simpo/metrics/aggregate_metrics.json +390 -0
  12. v1-undetermined/assembled_activations/llama32-3b-simpo/metrics/training_metrics.json +524 -0
  13. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/aligned_decoder_norm_heatmap.png +3 -0
  14. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/base_decoder_norm_heatmap.png +3 -0
  15. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/cf_shift_by_layer.png +3 -0
  16. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/cf_shift_p95_by_layer.png +3 -0
  17. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/class_distribution_multilayer.png +3 -0
  18. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/class_distribution_primary.png +3 -0
  19. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/cross_layer_cosine_drift_by_stream.png +3 -0
  20. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/decoder_norm_ratio_by_layer.png +3 -0
  21. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/feature_layer_trajectories.png +3 -0
  22. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/feature_sharing_ratio_by_layer.png +3 -0
  23. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/fve_by_layer.png +3 -0
  24. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/l0_by_layer.png +3 -0
  25. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/layer_concentration_entropy.png +3 -0
  26. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/loss_curves.png +3 -0
  27. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/max_norm_layer_migration.png +3 -0
  28. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/rho_histogram_by_layer.png +3 -0
  29. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/rho_theta_scatter_by_layer.png +3 -0
  30. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/superposition_by_layer.png +3 -0
  31. v1-undetermined/assembled_activations/llama32-3b-simpo/plots/theta_by_layer.png +3 -0
  32. v1-undetermined/assembled_activations/qwen3-4b-dpo/checkpoints/final.pt +3 -0
  33. v1-undetermined/assembled_activations/qwen3-4b-dpo/features/counterfactual_scores.csv +0 -0
  34. v1-undetermined/assembled_activations/qwen3-4b-dpo/features/counterfactual_scores_by_layer.csv +0 -0
  35. v1-undetermined/assembled_activations/qwen3-4b-dpo/features/cross_layer_cosine_drift.csv +3 -0
  36. v1-undetermined/assembled_activations/qwen3-4b-dpo/features/decoder_layer_profiles.csv +0 -0
  37. v1-undetermined/assembled_activations/qwen3-4b-dpo/features/feature_classification.csv +0 -0
  38. v1-undetermined/assembled_activations/qwen3-4b-dpo/features/merged_classification.csv +0 -0
  39. v1-undetermined/assembled_activations/qwen3-4b-dpo/features/model_layer_stream_patterns.csv +0 -0
  40. v1-undetermined/assembled_activations/qwen3-4b-dpo/features/superposition_analysis.json +0 -0
  41. v1-undetermined/assembled_activations/qwen3-4b-dpo/metrics/aggregate_metrics.json +375 -0
  42. v1-undetermined/assembled_activations/qwen3-4b-dpo/metrics/training_metrics.json +524 -0
  43. v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/aligned_decoder_norm_heatmap.png +3 -0
  44. v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/base_decoder_norm_heatmap.png +3 -0
  45. v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/cf_shift_by_layer.png +3 -0
  46. v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/cf_shift_p95_by_layer.png +3 -0
  47. v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/class_distribution_multilayer.png +3 -0
  48. v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/class_distribution_primary.png +3 -0
  49. v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/cross_layer_cosine_drift_by_stream.png +3 -0
  50. v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/decoder_norm_ratio_by_layer.png +3 -0
.gitattributes CHANGED
@@ -324,3 +324,12 @@ v1-undetermined/assembled_activations/llama32-3b-orpo/features/decoder_layer_pro
324
  v1-undetermined/assembled_activations/llama32-3b-ppo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
325
  v1-undetermined/assembled_activations/llama32-3b-ppo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
326
  v1-undetermined/assembled_activations/llama32-3b-ppo/features/model_layer_stream_patterns.csv filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
324
  v1-undetermined/assembled_activations/llama32-3b-ppo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
325
  v1-undetermined/assembled_activations/llama32-3b-ppo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
326
  v1-undetermined/assembled_activations/llama32-3b-ppo/features/model_layer_stream_patterns.csv filter=lfs diff=lfs merge=lfs -text
327
+ v1-undetermined/assembled_activations/llama32-3b-simpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
328
+ v1-undetermined/assembled_activations/llama32-3b-simpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
329
+ v1-undetermined/assembled_activations/llama32-3b-simpo/features/model_layer_stream_patterns.csv filter=lfs diff=lfs merge=lfs -text
330
+ v1-undetermined/assembled_activations/qwen3-4b-dpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
331
+ v1-undetermined/assembled_activations/qwen3-4b-grpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
332
+ v1-undetermined/assembled_activations/qwen3-4b-grpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
333
+ v1-undetermined/assembled_activations/qwen3-4b-kto/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
334
+ v1-undetermined/assembled_activations/qwen3-4b-kto/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
335
+ v1-undetermined/assembled_activations/qwen3-4b-orpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
v1-undetermined/assembled_activations/llama32-3b-simpo/checkpoints/final.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ccb6d15d31851a2c8616da9042c402d47e5863ff5af099931d9f1041b86f0d4
3
+ size 10873439922
v1-undetermined/assembled_activations/llama32-3b-simpo/features/counterfactual_scores.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-simpo/features/counterfactual_scores_by_layer.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-simpo/features/cross_layer_cosine_drift.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fde65a9d5d2e756fabc40ffd1a57e4cc6bfe199673e000347a8b295cc77d0e3
3
+ size 27349788
v1-undetermined/assembled_activations/llama32-3b-simpo/features/decoder_layer_profiles.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:419a4021e282db8cb1143e2576e64cb82eb8327da1cb2627676f9cdd475aa2a1
3
+ size 13236337
v1-undetermined/assembled_activations/llama32-3b-simpo/features/feature_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-simpo/features/merged_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-simpo/features/model_layer_stream_patterns.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98183a1f8d1d6f341e6de2278123c273f14d27872b9ead0fd01199623ec143e5
3
+ size 10829188
v1-undetermined/assembled_activations/llama32-3b-simpo/features/superposition_analysis.json ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-simpo/metrics/aggregate_metrics.json ADDED
@@ -0,0 +1,390 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crosscoder_kind": "multilayer_sparc",
3
+ "layers": [
4
+ 10,
5
+ 11,
6
+ 12
7
+ ],
8
+ "topk_mode": "model_balanced_layer_agg",
9
+ "topk": 400,
10
+ "expansion_factor": 8,
11
+ "dict_size": 24576,
12
+ "forced_shared_fraction": 0.06,
13
+ "class_counts": {
14
+ "aligned_only": 7032,
15
+ "shared_intermediate": 6309,
16
+ "base_only": 4631,
17
+ "shared_redirected": 3294,
18
+ "shared_attenuated": 1746,
19
+ "shared_aligned": 1564
20
+ },
21
+ "multilayer_class_counts": {
22
+ "drifting_or_rotating": 9101,
23
+ "persistent_aligned_only": 4366,
24
+ "persistent_shared": 3358,
25
+ "localized_aligned_only": 2666,
26
+ "persistent_base_only": 2377,
27
+ "localized_base_only": 2254,
28
+ "mixed_or_ambiguous": 454
29
+ },
30
+ "classification_thresholds": {
31
+ "rho_base_only": 0.4,
32
+ "rho_aligned_only": 0.5646282217167712,
33
+ "rho_shared_low": 0.43133134923386174,
34
+ "rho_shared_high": 0.5646282217167712
35
+ },
36
+ "threshold_sensitivity": {
37
+ "original": {
38
+ "aligned_only": 7032,
39
+ "shared_intermediate": 6309,
40
+ "base_only": 4631,
41
+ "shared_redirected": 3294,
42
+ "shared_attenuated": 1746,
43
+ "shared_aligned": 1564
44
+ },
45
+ "perturbed": {
46
+ "delta_-0.05": {
47
+ "base_only": 1938,
48
+ "aligned_only": 4114,
49
+ "shared_aligned": 1612,
50
+ "shared_redirected": 5669,
51
+ "shared_intermediate": 9535,
52
+ "shared_attenuated": 1708,
53
+ "other": 0
54
+ },
55
+ "delta_+0.05": {
56
+ "base_only": 7615,
57
+ "aligned_only": 10576,
58
+ "shared_aligned": 1486,
59
+ "shared_redirected": 872,
60
+ "shared_intermediate": 1690,
61
+ "shared_attenuated": 2337,
62
+ "other": 0
63
+ }
64
+ },
65
+ "perturbation": 0.05
66
+ },
67
+ "class_counts_by_layer": {
68
+ "10": {
69
+ "shared_intermediate": 7093,
70
+ "shared_redirected": 5890,
71
+ "base_only": 5623,
72
+ "shared_aligned": 3521,
73
+ "aligned_only": 2449
74
+ },
75
+ "11": {
76
+ "shared_attenuated": 8771,
77
+ "other": 5527,
78
+ "shared_intermediate": 4647,
79
+ "shared_redirected": 2565,
80
+ "shared_aligned": 1389,
81
+ "base_only": 1060,
82
+ "aligned_only": 617
83
+ },
84
+ "12": {
85
+ "shared_redirected": 9538,
86
+ "shared_intermediate": 5957,
87
+ "base_only": 3856,
88
+ "shared_aligned": 3719,
89
+ "aligned_only": 1506
90
+ }
91
+ },
92
+ "feature_sharing_ratio_by_layer": {
93
+ "10": 0.6715494791666666,
94
+ "11": 0.7068684895833334,
95
+ "12": 0.7818196614583334
96
+ },
97
+ "decoder_amplification_by_layer": {
98
+ "10": {
99
+ "median": 0.9999999870874021,
100
+ "p95": 5.846633247470493
101
+ },
102
+ "11": {
103
+ "median": 0.9999999890566127,
104
+ "p95": 2.8997828111006294
105
+ },
106
+ "12": {
107
+ "median": 0.999999988812232,
108
+ "p95": 3.348760753383095
109
+ }
110
+ },
111
+ "classification_thresholds_by_layer": {
112
+ "10": {
113
+ "rho_base_only": 0.3298648264878633,
114
+ "rho_aligned_only": 0.7696755668389346,
115
+ "rho_shared_low": 0.3298648264878633,
116
+ "rho_shared_high": 0.7696755668389346
117
+ },
118
+ "11": {
119
+ "rho_base_only": 0.2341652512550354,
120
+ "rho_aligned_only": 0.8046609088778496,
121
+ "rho_shared_low": 0.46459993720054626,
122
+ "rho_shared_high": 0.6093218177556992
123
+ },
124
+ "12": {
125
+ "rho_base_only": 0.33508244103501,
126
+ "rho_aligned_only": 0.7473176242691034,
127
+ "rho_shared_low": 0.33508244103501,
128
+ "rho_shared_high": 0.7473176242691034
129
+ }
130
+ },
131
+ "threshold_sensitivity_by_layer": {
132
+ "10": {
133
+ "original": {
134
+ "shared_intermediate": 7093,
135
+ "shared_redirected": 5890,
136
+ "base_only": 5623,
137
+ "shared_aligned": 3521,
138
+ "aligned_only": 2449
139
+ },
140
+ "perturbed": {
141
+ "delta_-0.05": {
142
+ "base_only": 3901,
143
+ "aligned_only": 1675,
144
+ "shared_aligned": 4473,
145
+ "shared_redirected": 6604,
146
+ "shared_intermediate": 7923,
147
+ "shared_attenuated": 0,
148
+ "other": 0
149
+ },
150
+ "delta_+0.05": {
151
+ "base_only": 7427,
152
+ "aligned_only": 3600,
153
+ "shared_aligned": 2604,
154
+ "shared_redirected": 4977,
155
+ "shared_intermediate": 5968,
156
+ "shared_attenuated": 0,
157
+ "other": 0
158
+ }
159
+ },
160
+ "perturbation": 0.05
161
+ },
162
+ "11": {
163
+ "original": {
164
+ "shared_attenuated": 8771,
165
+ "other": 5527,
166
+ "shared_intermediate": 4647,
167
+ "shared_redirected": 2565,
168
+ "shared_aligned": 1389,
169
+ "base_only": 1060,
170
+ "aligned_only": 617
171
+ },
172
+ "perturbed": {
173
+ "delta_-0.05": {
174
+ "base_only": 637,
175
+ "aligned_only": 374,
176
+ "shared_aligned": 2466,
177
+ "shared_redirected": 3929,
178
+ "shared_intermediate": 7325,
179
+ "shared_attenuated": 6633,
180
+ "other": 3212
181
+ },
182
+ "delta_+0.05": {
183
+ "base_only": 1818,
184
+ "aligned_only": 1085,
185
+ "shared_aligned": 54,
186
+ "shared_redirected": 814,
187
+ "shared_intermediate": 1509,
188
+ "shared_attenuated": 11833,
189
+ "other": 7463
190
+ }
191
+ },
192
+ "perturbation": 0.05
193
+ },
194
+ "12": {
195
+ "original": {
196
+ "shared_redirected": 9538,
197
+ "shared_intermediate": 5957,
198
+ "base_only": 3856,
199
+ "shared_aligned": 3719,
200
+ "aligned_only": 1506
201
+ },
202
+ "perturbed": {
203
+ "delta_-0.05": {
204
+ "base_only": 2386,
205
+ "aligned_only": 972,
206
+ "shared_aligned": 4290,
207
+ "shared_redirected": 10373,
208
+ "shared_intermediate": 6555,
209
+ "shared_attenuated": 0,
210
+ "other": 0
211
+ },
212
+ "delta_+0.05": {
213
+ "base_only": 6109,
214
+ "aligned_only": 2494,
215
+ "shared_aligned": 2781,
216
+ "shared_redirected": 8206,
217
+ "shared_intermediate": 4986,
218
+ "shared_attenuated": 0,
219
+ "other": 0
220
+ }
221
+ },
222
+ "perturbation": 0.05
223
+ }
224
+ },
225
+ "counterfactual_shift_by_layer": {
226
+ "10": {
227
+ "aligned_only": {
228
+ "mean_shift": 3.2320919899215418e-06,
229
+ "median_shift": 0.0,
230
+ "p95_abs_shift": 0.0,
231
+ "count": 7032
232
+ },
233
+ "base_only": {
234
+ "mean_shift": -3.5814365136129003e-06,
235
+ "median_shift": 0.0,
236
+ "p95_abs_shift": 0.0,
237
+ "count": 4631
238
+ },
239
+ "shared_aligned": {
240
+ "mean_shift": 5.724467744014439e-05,
241
+ "median_shift": 0.0,
242
+ "p95_abs_shift": 0.07152953743934631,
243
+ "count": 1564
244
+ },
245
+ "shared_attenuated": {
246
+ "mean_shift": -4.0716883922622904e-07,
247
+ "median_shift": 0.0,
248
+ "p95_abs_shift": 0.0,
249
+ "count": 1746
250
+ },
251
+ "shared_intermediate": {
252
+ "mean_shift": -7.30478225800923e-08,
253
+ "median_shift": 0.0,
254
+ "p95_abs_shift": 0.0,
255
+ "count": 6309
256
+ },
257
+ "shared_redirected": {
258
+ "mean_shift": 4.9160252745138344e-08,
259
+ "median_shift": 0.0,
260
+ "p95_abs_shift": 0.0,
261
+ "count": 3294
262
+ }
263
+ },
264
+ "11": {
265
+ "aligned_only": {
266
+ "mean_shift": 1.867720686331114e-05,
267
+ "median_shift": 0.0,
268
+ "p95_abs_shift": 0.0,
269
+ "count": 7032
270
+ },
271
+ "base_only": {
272
+ "mean_shift": -1.9894835705085598e-05,
273
+ "median_shift": 0.0,
274
+ "p95_abs_shift": 0.0,
275
+ "count": 4631
276
+ },
277
+ "shared_aligned": {
278
+ "mean_shift": 7.235741871226919e-05,
279
+ "median_shift": 0.0,
280
+ "p95_abs_shift": 0.0822630167007446,
281
+ "count": 1564
282
+ },
283
+ "shared_attenuated": {
284
+ "mean_shift": -3.471290511984925e-06,
285
+ "median_shift": 0.0,
286
+ "p95_abs_shift": 0.0,
287
+ "count": 1746
288
+ },
289
+ "shared_intermediate": {
290
+ "mean_shift": 8.675455243721306e-08,
291
+ "median_shift": 0.0,
292
+ "p95_abs_shift": 0.0,
293
+ "count": 6309
294
+ },
295
+ "shared_redirected": {
296
+ "mean_shift": -3.6504596481600716e-07,
297
+ "median_shift": 0.0,
298
+ "p95_abs_shift": 0.0,
299
+ "count": 3294
300
+ }
301
+ },
302
+ "12": {
303
+ "aligned_only": {
304
+ "mean_shift": 1.7178416289861293e-05,
305
+ "median_shift": 0.0,
306
+ "p95_abs_shift": 0.0,
307
+ "count": 7032
308
+ },
309
+ "base_only": {
310
+ "mean_shift": -1.7390496993661782e-05,
311
+ "median_shift": 0.0,
312
+ "p95_abs_shift": 0.0,
313
+ "count": 4631
314
+ },
315
+ "shared_aligned": {
316
+ "mean_shift": 0.00016059010712473373,
317
+ "median_shift": 0.0,
318
+ "p95_abs_shift": 0.08721660748124122,
319
+ "count": 1564
320
+ },
321
+ "shared_attenuated": {
322
+ "mean_shift": -4.775544404814352e-06,
323
+ "median_shift": 0.0,
324
+ "p95_abs_shift": 0.0,
325
+ "count": 1746
326
+ },
327
+ "shared_intermediate": {
328
+ "mean_shift": -4.4374906589709826e-08,
329
+ "median_shift": 0.0,
330
+ "p95_abs_shift": 0.0,
331
+ "count": 6309
332
+ },
333
+ "shared_redirected": {
334
+ "mean_shift": 3.799928669689255e-07,
335
+ "median_shift": 0.0,
336
+ "p95_abs_shift": 0.0,
337
+ "count": 3294
338
+ }
339
+ }
340
+ },
341
+ "total_features": 24576,
342
+ "fve_base": 0.6715279371326507,
343
+ "fve_aligned": 0.662245636523082,
344
+ "fve_base_by_layer": [
345
+ 0.6305625454293495,
346
+ 0.6942680309580259,
347
+ 0.6897531785265938
348
+ ],
349
+ "fve_aligned_by_layer": [
350
+ 0.6188769243774613,
351
+ 0.6861868694190579,
352
+ 0.6816730599128763
353
+ ],
354
+ "val_fve_base_by_layer": [
355
+ 0.6305625454293495,
356
+ 0.6942680309580259,
357
+ 0.6897531785265938
358
+ ],
359
+ "val_fve_aligned_by_layer": [
360
+ 0.6188769243774613,
361
+ 0.6861868694190579,
362
+ 0.6816730599128763
363
+ ],
364
+ "dead_neuron_fraction": 0.9593874006958921,
365
+ "l0_sparsity_base": 64.82328354345033,
366
+ "l0_sparsity_aligned": 63.5290322557083,
367
+ "l0_base_by_layer": [
368
+ 47.34114127262113,
369
+ 68.14484821949796,
370
+ 78.9838550788091
371
+ ],
372
+ "l0_aligned_by_layer": [
373
+ 46.07039915353182,
374
+ 66.73431115002919,
375
+ 77.78238105662581
376
+ ],
377
+ "val_l0_base": 66.31130234483649,
378
+ "val_l0_aligned": 64.88672773750665,
379
+ "val_l0_base_by_layer": [
380
+ 48.44813481675393,
381
+ 69.85569371727749,
382
+ 80.6300719895288
383
+ ],
384
+ "val_l0_aligned_by_layer": [
385
+ 46.990892240514306,
386
+ 68.39632419765931,
387
+ 79.27296032331377
388
+ ],
389
+ "superposition_fraction": 0.0
390
+ }
v1-undetermined/assembled_activations/llama32-3b-simpo/metrics/training_metrics.json ADDED
@@ -0,0 +1,524 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epochs": [
3
+ 1,
4
+ 2,
5
+ 3,
6
+ 4,
7
+ 5,
8
+ 6,
9
+ 7,
10
+ 8
11
+ ],
12
+ "train_loss": [
13
+ 0.05927932382030092,
14
+ 0.03255366562918412,
15
+ 0.03115293783746264,
16
+ 0.030371403902273254,
17
+ 0.02979045822211719,
18
+ 0.02931892105268931,
19
+ 0.028975081874163124,
20
+ 0.02878278101344799
21
+ ],
22
+ "val_loss": [
23
+ 0.033846099475759485,
24
+ 0.032007066728010854,
25
+ 0.031248764461637792,
26
+ 0.030814522370227967,
27
+ 0.030503301778856996,
28
+ 0.030317795097203778,
29
+ 0.030241612618038167,
30
+ 0.030229815228291206
31
+ ],
32
+ "train_fve_base": [
33
+ 0.503612844223043,
34
+ 0.6349237019808336,
35
+ 0.6552128129303282,
36
+ 0.6672707846550238,
37
+ 0.6764131289647465,
38
+ 0.6837705131587548,
39
+ 0.688966876536312,
40
+ 0.6917263760388559
41
+ ],
42
+ "train_fve_aligned": [
43
+ 0.49909401890992744,
44
+ 0.6284301312639222,
45
+ 0.6463590448736379,
46
+ 0.6569368737661135,
47
+ 0.6652233918299121,
48
+ 0.6720354633935348,
49
+ 0.676885349633606,
50
+ 0.6794380337191268
51
+ ],
52
+ "val_fve_base": [
53
+ 0.6182141828287334,
54
+ 0.6435623983438102,
55
+ 0.6547190727987838,
56
+ 0.6613033850779708,
57
+ 0.6664242657067264,
58
+ 0.6696532880448546,
59
+ 0.671212948429647,
60
+ 0.6715279371326507
61
+ ],
62
+ "val_fve_aligned": [
63
+ 0.613135617752974,
64
+ 0.6363836629852575,
65
+ 0.6465314929397943,
66
+ 0.6526829559765561,
67
+ 0.6575311690724958,
68
+ 0.6605612236167748,
69
+ 0.6619636584950991,
70
+ 0.662245636523082
71
+ ],
72
+ "val_fve_base_by_layer": [
73
+ [
74
+ 0.5780349083595875,
75
+ 0.6430585827502905,
76
+ 0.633548992466552
77
+ ],
78
+ [
79
+ 0.6026088904335861,
80
+ 0.6670344584275291,
81
+ 0.6610437787640158
82
+ ],
83
+ [
84
+ 0.6138415374056831,
85
+ 0.6778382768181606,
86
+ 0.6724773370782743
87
+ ],
88
+ [
89
+ 0.6206396946108154,
90
+ 0.6839593848632892,
91
+ 0.6793110280136787
92
+ ],
93
+ [
94
+ 0.6254869090949053,
95
+ 0.6891793920731669,
96
+ 0.6846064441491172
97
+ ],
98
+ [
99
+ 0.6288681657526506,
100
+ 0.6923003037562545,
101
+ 0.6877913343968816
102
+ ],
103
+ [
104
+ 0.6302341993566583,
105
+ 0.6938901640357772,
106
+ 0.6895144166746688
107
+ ],
108
+ [
109
+ 0.6305625454293495,
110
+ 0.6942680309580259,
111
+ 0.6897531785265938
112
+ ]
113
+ ],
114
+ "val_fve_aligned_by_layer": [
115
+ [
116
+ 0.5722053659523969,
117
+ 0.6384917887717642,
118
+ 0.6287096414266456
119
+ ],
120
+ [
121
+ 0.5942014565018459,
122
+ 0.6606073991166359,
123
+ 0.6543420678033879
124
+ ],
125
+ [
126
+ 0.6036381100484838,
127
+ 0.6709235947169558,
128
+ 0.6650327238112844
129
+ ],
130
+ [
131
+ 0.609793980708297,
132
+ 0.676685466816288,
133
+ 0.6715693632969681
134
+ ],
135
+ [
136
+ 0.6141581582149286,
137
+ 0.6816896346851169,
138
+ 0.6767456603299885
139
+ ],
140
+ [
141
+ 0.6174167641170362,
142
+ 0.6844394194518084,
143
+ 0.6798274301733646
144
+ ],
145
+ [
146
+ 0.6185780351698711,
147
+ 0.6858815141997412,
148
+ 0.6814313761850926
149
+ ],
150
+ [
151
+ 0.6188769243774613,
152
+ 0.6861868694190579,
153
+ 0.6816730599128763
154
+ ]
155
+ ],
156
+ "train_fve_base_by_layer": [
157
+ [
158
+ 0.4677472282277167,
159
+ 0.5262350366258928,
160
+ 0.5168562211700858
161
+ ],
162
+ [
163
+ 0.5941825219920972,
164
+ 0.6585408136709628,
165
+ 0.6520477089350222
166
+ ],
167
+ [
168
+ 0.6142003669491026,
169
+ 0.6779960200094857,
170
+ 0.6734419917406148
171
+ ],
172
+ [
173
+ 0.6260286404664797,
174
+ 0.6899296233894807,
175
+ 0.6858540300869204
176
+ ],
177
+ [
178
+ 0.6346775183744481,
179
+ 0.6991928598369971,
180
+ 0.6953689465032846
181
+ ],
182
+ [
183
+ 0.6413471689714163,
184
+ 0.7068967281623814,
185
+ 0.7030675794670475
186
+ ],
187
+ [
188
+ 0.645974550411427,
189
+ 0.7123980270932448,
190
+ 0.7085279913861642
191
+ ],
192
+ [
193
+ 0.6484792617699311,
194
+ 0.7152978126514188,
195
+ 0.7114019931510952
196
+ ]
197
+ ],
198
+ "train_fve_aligned_by_layer": [
199
+ [
200
+ 0.4632670403361947,
201
+ 0.5220091638687544,
202
+ 0.5120058040554872
203
+ ],
204
+ [
205
+ 0.5868167219214876,
206
+ 0.6527098008087345,
207
+ 0.6457638119440363
208
+ ],
209
+ [
210
+ 0.6038263409164171,
211
+ 0.6701786429716423,
212
+ 0.6650720916153462
213
+ ],
214
+ [
215
+ 0.6137546507921124,
216
+ 0.6809061225297085,
217
+ 0.6761497881978976
218
+ ],
219
+ [
220
+ 0.6210833272249185,
221
+ 0.6895746217091322,
222
+ 0.685012164898108
223
+ ],
224
+ [
225
+ 0.6268049689043493,
226
+ 0.6969297279479276,
227
+ 0.692371636472525
228
+ ],
229
+ [
230
+ 0.6309437568745833,
231
+ 0.7021611605689878,
232
+ 0.6975510722353524
233
+ ],
234
+ [
235
+ 0.6331689965634477,
236
+ 0.7048862822836627,
237
+ 0.7002587634363302
238
+ ]
239
+ ],
240
+ "dead_neurons": [
241
+ 0.8715425254097475,
242
+ 0.9557898779118611,
243
+ 0.961711463702748,
244
+ 0.9616781369817236,
245
+ 0.9606253722224817,
246
+ 0.9598153939876234,
247
+ 0.9594147409254253,
248
+ 0.9593874006958921
249
+ ],
250
+ "l0_base": [
251
+ 110.7568565675966,
252
+ 59.56439248623378,
253
+ 59.25963353136444,
254
+ 60.706109090917515,
255
+ 62.3549953445105,
256
+ 63.71940258268856,
257
+ 64.55443870290566,
258
+ 64.82328354345033
259
+ ],
260
+ "l0_aligned": [
261
+ 109.46531757414863,
262
+ 58.278739833442074,
263
+ 58.12394318856207,
264
+ 59.642915827032425,
265
+ 61.2593235343225,
266
+ 62.529463801965754,
267
+ 63.29711094114144,
268
+ 63.5290322557083
269
+ ],
270
+ "l0_base_by_layer": [
271
+ [
272
+ 93.83470154699359,
273
+ 118.0552393461763,
274
+ 120.38061879743141
275
+ ],
276
+ [
277
+ 43.45864346176299,
278
+ 63.35460449503795,
279
+ 71.87992556917688
280
+ ],
281
+ [
282
+ 43.179728546409805,
283
+ 62.33630691768827,
284
+ 72.2628612084063
285
+ ],
286
+ [
287
+ 44.387076765907764,
288
+ 63.711981903093985,
289
+ 74.01926444833626
290
+ ],
291
+ [
292
+ 45.60799766491535,
293
+ 65.49543928779919,
294
+ 75.96154407472271
295
+ ],
296
+ [
297
+ 46.5932209573847,
298
+ 66.9508720081728,
299
+ 77.61410901926445
300
+ ],
301
+ [
302
+ 47.15878575598366,
303
+ 67.86980078809107,
304
+ 78.63472343841214
305
+ ],
306
+ [
307
+ 47.34114127262113,
308
+ 68.14484821949796,
309
+ 78.9838550788091
310
+ ]
311
+ ],
312
+ "l0_aligned_by_layer": [
313
+ [
314
+ 92.87680604203152,
315
+ 116.69826328079392,
316
+ 118.8208734676007
317
+ ],
318
+ [
319
+ 42.17456582019848,
320
+ 62.11866973146527,
321
+ 70.5429801517805
322
+ ],
323
+ [
324
+ 41.8616097489784,
325
+ 61.297577349678924,
326
+ 71.2126386456509
327
+ ],
328
+ [
329
+ 43.14568739054291,
330
+ 62.718056771745474,
331
+ 73.06499927028605
332
+ ],
333
+ [
334
+ 44.36987375948628,
335
+ 64.42671847635727,
336
+ 74.98137405137186
337
+ ],
338
+ [
339
+ 45.319158639813196,
340
+ 65.73668272037361,
341
+ 76.53254524226503
342
+ ],
343
+ [
344
+ 45.89940893169877,
345
+ 66.51408347927612,
346
+ 77.47783493870404
347
+ ],
348
+ [
349
+ 46.07039915353182,
350
+ 66.73431115002919,
351
+ 77.78238105662581
352
+ ]
353
+ ],
354
+ "val_l0_base": [
355
+ 62.84433352884822,
356
+ 60.29535830832277,
357
+ 61.20708046159195,
358
+ 62.52763392663127,
359
+ 64.26254554069479,
360
+ 65.39581746206234,
361
+ 66.14812595307515,
362
+ 66.31130234483649
363
+ ],
364
+ "val_l0_aligned": [
365
+ 61.65801482674963,
366
+ 58.96664246963581,
367
+ 59.968006054144254,
368
+ 61.42377619718382,
369
+ 63.13096464866119,
370
+ 64.15648820887061,
371
+ 64.74734806639987,
372
+ 64.88672773750665
373
+ ],
374
+ "val_l0_base_by_layer": [
375
+ [
376
+ 46.69628055432704,
377
+ 67.60760254385583,
378
+ 74.2291121557745
379
+ ],
380
+ [
381
+ 44.08933246073298,
382
+ 63.24301920386509,
383
+ 73.55371948561744
384
+ ],
385
+ [
386
+ 44.62358203109022,
387
+ 64.35722079451796,
388
+ 74.6404341453033
389
+ ],
390
+ [
391
+ 45.78910340314136,
392
+ 65.5068172035417,
393
+ 76.28697643979058
394
+ ],
395
+ [
396
+ 46.838787098829656,
397
+ 67.65003272251309,
398
+ 78.29881110865408
399
+ ],
400
+ [
401
+ 47.84876746532181,
402
+ 68.77879581151832,
403
+ 79.55988219895288
404
+ ],
405
+ [
406
+ 48.298047570033845,
407
+ 69.62587262697869,
408
+ 80.52045157068063
409
+ ],
410
+ [
411
+ 48.44813481675393,
412
+ 69.85569371727749,
413
+ 80.6300719895288
414
+ ]
415
+ ],
416
+ "val_l0_aligned_by_layer": [
417
+ [
418
+ 45.42937391590698,
419
+ 66.43193717277487,
420
+ 73.11272905759162
421
+ ],
422
+ [
423
+ 42.52116056512163,
424
+ 62.154504805959334,
425
+ 72.22425830301815
426
+ ],
427
+ [
428
+ 43.128981245749905,
429
+ 63.41393979057592,
430
+ 73.36109293193718
431
+ ],
432
+ [
433
+ 44.41513962271326,
434
+ 64.71924083769633,
435
+ 75.13694371727749
436
+ ],
437
+ [
438
+ 45.48292976898673,
439
+ 66.78866712959649,
440
+ 77.12129147514622
441
+ ],
442
+ [
443
+ 46.53888526636893,
444
+ 67.60274869109948,
445
+ 78.32782505694485
446
+ ],
447
+ [
448
+ 46.87685428239912,
449
+ 68.23958335996298,
450
+ 79.12559992605479
451
+ ],
452
+ [
453
+ 46.990892240514306,
454
+ 68.39632419765931,
455
+ 79.27296032331377
456
+ ]
457
+ ],
458
+ "self_recon": [
459
+ 0.027993987802083356,
460
+ 0.020580280668407207,
461
+ 0.019489796229263563,
462
+ 0.018846164809260463,
463
+ 0.018348425166802105,
464
+ 0.017941102601361603,
465
+ 0.017650670092498533,
466
+ 0.01749732045938467
467
+ ],
468
+ "cross_recon": [
469
+ 0.028028677605305648,
470
+ 0.02058453171394909,
471
+ 0.019495135330868548,
472
+ 0.018860481901260683,
473
+ 0.018386862897496265,
474
+ 0.018007030048376626,
475
+ 0.01773270223207089,
476
+ 0.017584875620731955
477
+ ],
478
+ "sparsity": [
479
+ 0.020073864794261775,
480
+ 0.0037395720159709976,
481
+ 0.00386508731269351,
482
+ 0.0039810461591020165,
483
+ 0.004087287744385409,
484
+ 0.004175006297253338,
485
+ 0.004231330718039396,
486
+ 0.004251510163788699
487
+ ],
488
+ "val_self_recon": [
489
+ 0.021518522912529127,
490
+ 0.020130771841288236,
491
+ 0.01952393066274558,
492
+ 0.019163148765359562,
493
+ 0.01887795397358415,
494
+ 0.018701545738112862,
495
+ 0.018615932642672386,
496
+ 0.018599209112373633
497
+ ],
498
+ "val_cross_recon": [
499
+ 0.021522300454412457,
500
+ 0.020133654884408905,
501
+ 0.019527446098274587,
502
+ 0.019169504778625453,
503
+ 0.018888392441793887,
504
+ 0.01871574230745669,
505
+ 0.01863224525979832,
506
+ 0.018615821108261016
507
+ ],
508
+ "val_sparsity": [
509
+ 0.0037186559794363397,
510
+ 0.003822832755959003,
511
+ 0.003913855225247629,
512
+ 0.003983571632955399,
513
+ 0.004069990641803174,
514
+ 0.004129952197670156,
515
+ 0.004172781662752179,
516
+ 0.004184277574604874
517
+ ],
518
+ "layers": [
519
+ 10,
520
+ 11,
521
+ 12
522
+ ],
523
+ "topk_mode": "model_balanced_layer_agg"
524
+ }
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/aligned_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: 6d148f8c3598a04c1a3c8a3614f9ac7d755dbc992e11380fde57b5b27a6160f4
  • Pointer size: 131 Bytes
  • Size of remote file: 181 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/base_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: 90378eba9ab620720e9a5be63ddf19d857f163754dae2a9c0e11761109eef681
  • Pointer size: 131 Bytes
  • Size of remote file: 180 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/cf_shift_by_layer.png ADDED

Git LFS Details

  • SHA256: d2b279e1177392223a89e84117484da01bcfca14b6e209d50dfe53e7b52ce725
  • Pointer size: 130 Bytes
  • Size of remote file: 43.7 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/cf_shift_p95_by_layer.png ADDED

Git LFS Details

  • SHA256: 851ff1df0a1a1053cce768efd353a568e5f95567794ab2f62307710fbf3079b4
  • Pointer size: 131 Bytes
  • Size of remote file: 125 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/class_distribution_multilayer.png ADDED

Git LFS Details

  • SHA256: ad3199c1872d52c1b76cad0bfbecc17df5b3e7214c28e517a1c98dc434feb566
  • Pointer size: 130 Bytes
  • Size of remote file: 87.1 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/class_distribution_primary.png ADDED

Git LFS Details

  • SHA256: 339f0f6aba2c93aae7ec7a30bc58f71b369d7f8067bf58a8f2d3e737fec8638b
  • Pointer size: 131 Bytes
  • Size of remote file: 148 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/cross_layer_cosine_drift_by_stream.png ADDED

Git LFS Details

  • SHA256: 533c9c7ee417e3884b79be016c7d37e19499171e41899c787d94f7087e672374
  • Pointer size: 130 Bytes
  • Size of remote file: 50.2 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/decoder_norm_ratio_by_layer.png ADDED

Git LFS Details

  • SHA256: 644a033382361decd1c16c9c403e22528a058b38708031dcf936d0fc13c72d9a
  • Pointer size: 130 Bytes
  • Size of remote file: 51.4 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/feature_layer_trajectories.png ADDED

Git LFS Details

  • SHA256: 73106568eb9aeee5a10abf229d4bed081a19f449bdd8ae642ee8b248f0f14f34
  • Pointer size: 131 Bytes
  • Size of remote file: 658 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/feature_sharing_ratio_by_layer.png ADDED

Git LFS Details

  • SHA256: 546e91aeafe39f3a08115604051f0e363f6a1d19c9c18cac97fd3ab2e2ae0e7c
  • Pointer size: 130 Bytes
  • Size of remote file: 55.3 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/fve_by_layer.png ADDED

Git LFS Details

  • SHA256: 29a2c6dc652c68c5db0866990841863f7fbb73177badb1253971d775b81d05a2
  • Pointer size: 131 Bytes
  • Size of remote file: 111 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/l0_by_layer.png ADDED

Git LFS Details

  • SHA256: 2fa4c740150bb2c6e395c42a3fbc42e81add40686df232319a3821d4afb4008f
  • Pointer size: 130 Bytes
  • Size of remote file: 93.7 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/layer_concentration_entropy.png ADDED

Git LFS Details

  • SHA256: 77c88abe809a20c6260529ec09c812d3fc7f327ceca3fa0e4d339e2339e3447d
  • Pointer size: 131 Bytes
  • Size of remote file: 106 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/loss_curves.png ADDED

Git LFS Details

  • SHA256: 7da65a8bfbcc7548869c6d28d685ec0edbe575c6a0080e4df97fc2599401aeb1
  • Pointer size: 131 Bytes
  • Size of remote file: 301 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/max_norm_layer_migration.png ADDED

Git LFS Details

  • SHA256: 90905d06c7c2260d8581e90c8c7361d50b4c1d04b09b391d04d1f6f4e790c24f
  • Pointer size: 130 Bytes
  • Size of remote file: 83 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/rho_histogram_by_layer.png ADDED

Git LFS Details

  • SHA256: db116820ff11ebc613106d1ed76dc035f2361577131214fbe84d8b0c208dfb91
  • Pointer size: 130 Bytes
  • Size of remote file: 48.5 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/rho_theta_scatter_by_layer.png ADDED

Git LFS Details

  • SHA256: 1ffbb671d608deaf20429fb7b7161cfe947d4dc7099bf184651c4109022125d3
  • Pointer size: 132 Bytes
  • Size of remote file: 1.44 MB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/superposition_by_layer.png ADDED

Git LFS Details

  • SHA256: f2c091fab892c52dd76137c531d910fceb852741343c66a11f6d205b1374da37
  • Pointer size: 131 Bytes
  • Size of remote file: 104 kB
v1-undetermined/assembled_activations/llama32-3b-simpo/plots/theta_by_layer.png ADDED

Git LFS Details

  • SHA256: 86bb23a3595d6f9caa4c9b260c2954a09e6f89e81dd88466fa6fb9397201b486
  • Pointer size: 130 Bytes
  • Size of remote file: 53.2 kB
v1-undetermined/assembled_activations/qwen3-4b-dpo/checkpoints/final.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c11ca07e546290482e8ffb66437612810b691e281b2b51e5fafa63d0183e6ba
3
+ size 7551253274
v1-undetermined/assembled_activations/qwen3-4b-dpo/features/counterfactual_scores.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/qwen3-4b-dpo/features/counterfactual_scores_by_layer.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/qwen3-4b-dpo/features/cross_layer_cosine_drift.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3febf3e5d588e9073d0bb305a5ed726655e667c74c78e38e55ff23802e330156
3
+ size 22223458
v1-undetermined/assembled_activations/qwen3-4b-dpo/features/decoder_layer_profiles.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/qwen3-4b-dpo/features/feature_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/qwen3-4b-dpo/features/merged_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/qwen3-4b-dpo/features/model_layer_stream_patterns.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/qwen3-4b-dpo/features/superposition_analysis.json ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/qwen3-4b-dpo/metrics/aggregate_metrics.json ADDED
@@ -0,0 +1,375 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crosscoder_kind": "multilayer_sparc",
3
+ "layers": [
4
+ 21,
5
+ 22,
6
+ 23
7
+ ],
8
+ "topk_mode": "model_balanced_layer_agg",
9
+ "topk": 400,
10
+ "expansion_factor": 8,
11
+ "dict_size": 20480,
12
+ "forced_shared_fraction": 0.06,
13
+ "class_counts": {
14
+ "shared_attenuated": 6613,
15
+ "shared_redirected": 6086,
16
+ "aligned_only": 5130,
17
+ "shared_intermediate": 1420,
18
+ "shared_aligned": 1230,
19
+ "base_only": 1
20
+ },
21
+ "multilayer_class_counts": {
22
+ "drifting_or_rotating": 12967,
23
+ "persistent_aligned_only": 5105,
24
+ "persistent_shared": 2374,
25
+ "localized_aligned_only": 25,
26
+ "mixed_or_ambiguous": 8,
27
+ "persistent_base_only": 1
28
+ },
29
+ "classification_thresholds": {
30
+ "rho_base_only": 0.4,
31
+ "rho_aligned_only": 0.5098805635251811,
32
+ "rho_shared_low": 0.4938351085309566,
33
+ "rho_shared_high": 0.5098805635251811
34
+ },
35
+ "threshold_sensitivity": {
36
+ "original": {
37
+ "shared_attenuated": 6613,
38
+ "shared_redirected": 6086,
39
+ "aligned_only": 5130,
40
+ "shared_intermediate": 1420,
41
+ "shared_aligned": 1230,
42
+ "base_only": 1
43
+ },
44
+ "perturbed": {
45
+ "delta_-0.05": {
46
+ "base_only": 0,
47
+ "aligned_only": 32,
48
+ "shared_aligned": 1272,
49
+ "shared_redirected": 15162,
50
+ "shared_intermediate": 3980,
51
+ "shared_attenuated": 34,
52
+ "other": 0
53
+ },
54
+ "delta_+0.05": {
55
+ "base_only": 58,
56
+ "aligned_only": 20333,
57
+ "shared_aligned": 0,
58
+ "shared_redirected": 0,
59
+ "shared_intermediate": 0,
60
+ "shared_attenuated": 89,
61
+ "other": 0
62
+ }
63
+ },
64
+ "perturbation": 0.05
65
+ },
66
+ "class_counts_by_layer": {
67
+ "21": {
68
+ "shared_attenuated": 6076,
69
+ "aligned_only": 5674,
70
+ "shared_redirected": 5184,
71
+ "shared_intermediate": 2053,
72
+ "shared_aligned": 1291,
73
+ "base_only": 202
74
+ },
75
+ "22": {
76
+ "base_only": 11110,
77
+ "aligned_only": 9370
78
+ },
79
+ "23": {
80
+ "base_only": 16899,
81
+ "aligned_only": 3581
82
+ }
83
+ },
84
+ "feature_sharing_ratio_by_layer": {
85
+ "21": 0.7130859375,
86
+ "22": 0.0,
87
+ "23": 0.0
88
+ },
89
+ "decoder_amplification_by_layer": {
90
+ "21": {
91
+ "median": 0.9999999872538781,
92
+ "p95": 1.2577608658612338
93
+ },
94
+ "22": {
95
+ "median": 0.9999999883079819,
96
+ "p95": 1.189090518979472
97
+ },
98
+ "23": {
99
+ "median": 0.9999999875603767,
100
+ "p95": 1.1453740443162634
101
+ }
102
+ },
103
+ "classification_thresholds_by_layer": {
104
+ "21": {
105
+ "rho_base_only": 0.4,
106
+ "rho_aligned_only": 0.5189358319923506,
107
+ "rho_shared_low": 0.48485935046201967,
108
+ "rho_shared_high": 0.5189358319923506
109
+ },
110
+ "22": {
111
+ "rho_base_only": 0.5011910310342542,
112
+ "rho_aligned_only": 0.5011910310342542,
113
+ "rho_shared_low": 0.5011910310342542,
114
+ "rho_shared_high": 0.5011910310342542
115
+ },
116
+ "23": {
117
+ "rho_base_only": 0.518398331752978,
118
+ "rho_aligned_only": 0.518398331752978,
119
+ "rho_shared_low": 0.518398331752978,
120
+ "rho_shared_high": 0.518398331752978
121
+ }
122
+ },
123
+ "threshold_sensitivity_by_layer": {
124
+ "21": {
125
+ "original": {
126
+ "shared_attenuated": 6076,
127
+ "aligned_only": 5674,
128
+ "shared_redirected": 5184,
129
+ "shared_intermediate": 2053,
130
+ "shared_aligned": 1291,
131
+ "base_only": 202
132
+ },
133
+ "perturbed": {
134
+ "delta_-0.05": {
135
+ "base_only": 88,
136
+ "aligned_only": 595,
137
+ "shared_aligned": 1421,
138
+ "shared_redirected": 12801,
139
+ "shared_intermediate": 4994,
140
+ "shared_attenuated": 581,
141
+ "other": 0
142
+ },
143
+ "delta_+0.05": {
144
+ "base_only": 1355,
145
+ "aligned_only": 17062,
146
+ "shared_aligned": 0,
147
+ "shared_redirected": 0,
148
+ "shared_intermediate": 0,
149
+ "shared_attenuated": 2063,
150
+ "other": 0
151
+ }
152
+ },
153
+ "perturbation": 0.05
154
+ },
155
+ "22": {
156
+ "original": {
157
+ "base_only": 11110,
158
+ "aligned_only": 9370
159
+ },
160
+ "perturbed": {
161
+ "delta_-0.05": {
162
+ "base_only": 674,
163
+ "aligned_only": 579,
164
+ "shared_aligned": 1370,
165
+ "shared_redirected": 13204,
166
+ "shared_intermediate": 4653,
167
+ "shared_attenuated": 0,
168
+ "other": 0
169
+ },
170
+ "delta_+0.05": {
171
+ "base_only": 19901,
172
+ "aligned_only": 579,
173
+ "shared_aligned": 0,
174
+ "shared_redirected": 0,
175
+ "shared_intermediate": 0,
176
+ "shared_attenuated": 0,
177
+ "other": 0
178
+ }
179
+ },
180
+ "perturbation": 0.05
181
+ },
182
+ "23": {
183
+ "original": {
184
+ "base_only": 16899,
185
+ "aligned_only": 3581
186
+ },
187
+ "perturbed": {
188
+ "delta_-0.05": {
189
+ "base_only": 1254,
190
+ "aligned_only": 33,
191
+ "shared_aligned": 1437,
192
+ "shared_redirected": 13401,
193
+ "shared_intermediate": 4355,
194
+ "shared_attenuated": 0,
195
+ "other": 0
196
+ },
197
+ "delta_+0.05": {
198
+ "base_only": 20447,
199
+ "aligned_only": 33,
200
+ "shared_aligned": 0,
201
+ "shared_redirected": 0,
202
+ "shared_intermediate": 0,
203
+ "shared_attenuated": 0,
204
+ "other": 0
205
+ }
206
+ },
207
+ "perturbation": 0.05
208
+ }
209
+ },
210
+ "counterfactual_shift_by_layer": {
211
+ "21": {
212
+ "aligned_only": {
213
+ "mean_shift": 4.2248451098280196e-05,
214
+ "median_shift": 0.0,
215
+ "p95_abs_shift": 0.0,
216
+ "count": 5130
217
+ },
218
+ "base_only": {
219
+ "mean_shift": -0.00044783533667214215,
220
+ "median_shift": -0.00044783533667214215,
221
+ "p95_abs_shift": 0.0,
222
+ "count": 1
223
+ },
224
+ "shared_aligned": {
225
+ "mean_shift": 0.0001278747351762846,
226
+ "median_shift": 0.0,
227
+ "p95_abs_shift": 0.40083945393562315,
228
+ "count": 1230
229
+ },
230
+ "shared_attenuated": {
231
+ "mean_shift": -2.484869482189629e-05,
232
+ "median_shift": 0.0,
233
+ "p95_abs_shift": 0.0,
234
+ "count": 6613
235
+ },
236
+ "shared_intermediate": {
237
+ "mean_shift": -1.8556327565504076e-05,
238
+ "median_shift": 0.0,
239
+ "p95_abs_shift": 0.0,
240
+ "count": 1420
241
+ },
242
+ "shared_redirected": {
243
+ "mean_shift": 9.42109240969424e-06,
244
+ "median_shift": 0.0,
245
+ "p95_abs_shift": 0.0,
246
+ "count": 6086
247
+ }
248
+ },
249
+ "22": {
250
+ "aligned_only": {
251
+ "mean_shift": 6.435667285540936e-05,
252
+ "median_shift": 0.0,
253
+ "p95_abs_shift": 0.0,
254
+ "count": 5130
255
+ },
256
+ "base_only": {
257
+ "mean_shift": -0.0016646130243316293,
258
+ "median_shift": -0.0016646130243316293,
259
+ "p95_abs_shift": 0.0,
260
+ "count": 1
261
+ },
262
+ "shared_aligned": {
263
+ "mean_shift": -6.292000267510558e-05,
264
+ "median_shift": 0.0,
265
+ "p95_abs_shift": 0.4378678083419799,
266
+ "count": 1230
267
+ },
268
+ "shared_attenuated": {
269
+ "mean_shift": -3.422924652860993e-05,
270
+ "median_shift": 0.0,
271
+ "p95_abs_shift": 0.0,
272
+ "count": 6613
273
+ },
274
+ "shared_intermediate": {
275
+ "mean_shift": -4.944617788833858e-05,
276
+ "median_shift": 0.0,
277
+ "p95_abs_shift": 0.0,
278
+ "count": 1420
279
+ },
280
+ "shared_redirected": {
281
+ "mean_shift": 6.49749713189548e-06,
282
+ "median_shift": 0.0,
283
+ "p95_abs_shift": 0.0,
284
+ "count": 6086
285
+ }
286
+ },
287
+ "23": {
288
+ "aligned_only": {
289
+ "mean_shift": 3.657554636571588e-05,
290
+ "median_shift": 0.0,
291
+ "p95_abs_shift": 0.0,
292
+ "count": 5130
293
+ },
294
+ "base_only": {
295
+ "mean_shift": -0.005122000351548195,
296
+ "median_shift": -0.005122000351548195,
297
+ "p95_abs_shift": 0.0,
298
+ "count": 1
299
+ },
300
+ "shared_aligned": {
301
+ "mean_shift": 0.0004172284026033713,
302
+ "median_shift": 0.0,
303
+ "p95_abs_shift": 0.46311920285224917,
304
+ "count": 1230
305
+ },
306
+ "shared_attenuated": {
307
+ "mean_shift": -5.05945687451926e-05,
308
+ "median_shift": 0.0,
309
+ "p95_abs_shift": 0.0,
310
+ "count": 6613
311
+ },
312
+ "shared_intermediate": {
313
+ "mean_shift": -4.6107667643326e-05,
314
+ "median_shift": 0.0,
315
+ "p95_abs_shift": 0.0,
316
+ "count": 1420
317
+ },
318
+ "shared_redirected": {
319
+ "mean_shift": 2.041490491879189e-05,
320
+ "median_shift": 0.0,
321
+ "p95_abs_shift": 0.0,
322
+ "count": 6086
323
+ }
324
+ }
325
+ },
326
+ "total_features": 20480,
327
+ "fve_base": 0.8757055364354118,
328
+ "fve_aligned": 0.8752784925605614,
329
+ "fve_base_by_layer": [
330
+ 0.873175848841043,
331
+ 0.8781832818585541,
332
+ 0.8757574155692655
333
+ ],
334
+ "fve_aligned_by_layer": [
335
+ 0.8727765308000655,
336
+ 0.8776650928077897,
337
+ 0.875393784795132
338
+ ],
339
+ "val_fve_base_by_layer": [
340
+ 0.873175848841043,
341
+ 0.8781832818585541,
342
+ 0.8757574155692655
343
+ ],
344
+ "val_fve_aligned_by_layer": [
345
+ 0.8727765308000655,
346
+ 0.8776650928077897,
347
+ 0.875393784795132
348
+ ],
349
+ "dead_neuron_fraction": 0.9698438135351414,
350
+ "l0_sparsity_base": 195.12645257792832,
351
+ "l0_sparsity_aligned": 195.01710472841813,
352
+ "l0_base_by_layer": [
353
+ 189.62868505545825,
354
+ 196.90103254524226,
355
+ 198.84962419731465
356
+ ],
357
+ "l0_aligned_by_layer": [
358
+ 189.5982742265032,
359
+ 196.52698117338002,
360
+ 198.92604349095154
361
+ ],
362
+ "val_l0_base": 196.35193550149808,
363
+ "val_l0_aligned": 196.22006401341622,
364
+ "val_l0_base_by_layer": [
365
+ 190.95555194635043,
366
+ 198.10874787175842,
367
+ 199.99149214659687
368
+ ],
369
+ "val_l0_aligned_by_layer": [
370
+ 190.91961174610398,
371
+ 197.67152053273784,
372
+ 200.0690445026178
373
+ ],
374
+ "superposition_fraction": 0.0
375
+ }
v1-undetermined/assembled_activations/qwen3-4b-dpo/metrics/training_metrics.json ADDED
@@ -0,0 +1,524 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epochs": [
3
+ 1,
4
+ 2,
5
+ 3,
6
+ 4,
7
+ 5,
8
+ 6,
9
+ 7,
10
+ 8
11
+ ],
12
+ "train_loss": [
13
+ 4.609393736596561,
14
+ 2.69254030662187,
15
+ 1.8937292124051763,
16
+ 1.490718465692038,
17
+ 1.2056996765988512,
18
+ 1.067367364353975,
19
+ 0.962318444314646,
20
+ 0.9284910069386025
21
+ ],
22
+ "val_loss": [
23
+ 3.842296364731814,
24
+ 2.838028597270007,
25
+ 1.675884915896111,
26
+ 1.5459347826024001,
27
+ 1.2952628450868018,
28
+ 1.1187260578440121,
29
+ 0.9881974970483031,
30
+ 0.9847213469874796
31
+ ],
32
+ "train_fve_base": [
33
+ 0.7009528921717278,
34
+ 0.8414981104933722,
35
+ 0.8591816361995726,
36
+ 0.8684247201531516,
37
+ 0.8748822911260286,
38
+ 0.8793395030101836,
39
+ 0.8825036513478312,
40
+ 0.8842055773568028
41
+ ],
42
+ "train_fve_aligned": [
43
+ 0.6976124361456679,
44
+ 0.8402781325408192,
45
+ 0.8584823833177191,
46
+ 0.8677803249089118,
47
+ 0.8742450237343761,
48
+ 0.8786993917149546,
49
+ 0.8818145939570314,
50
+ 0.8834975749793557
51
+ ],
52
+ "val_fve_base": [
53
+ 0.8268243592447011,
54
+ 0.8509355324725206,
55
+ 0.8617972332145531,
56
+ 0.8673466431532855,
57
+ 0.871497017550843,
58
+ 0.8741527851963542,
59
+ 0.8754691940327589,
60
+ 0.8757055364354118
61
+ ],
62
+ "val_fve_aligned": [
63
+ 0.8248398600448489,
64
+ 0.8503238387132814,
65
+ 0.8611327078954087,
66
+ 0.8668939870689552,
67
+ 0.8712653704962805,
68
+ 0.873785453317053,
69
+ 0.8750437828882827,
70
+ 0.8752784925605614
71
+ ],
72
+ "val_fve_base_by_layer": [
73
+ [
74
+ 0.8251121321897856,
75
+ 0.8281116719645355,
76
+ 0.8272492264577855
77
+ ],
78
+ [
79
+ 0.8484343571188562,
80
+ 0.8526254174596977,
81
+ 0.8517467794618058
82
+ ],
83
+ [
84
+ 0.8590603832799103,
85
+ 0.8641286879309809,
86
+ 0.8622025685160571
87
+ ],
88
+ [
89
+ 0.8647310518469485,
90
+ 0.8696209239085931,
91
+ 0.8676878922272727
92
+ ],
93
+ [
94
+ 0.8685201574994631,
95
+ 0.8741588498909436,
96
+ 0.8718119850333449
97
+ ],
98
+ [
99
+ 0.8714880250511369,
100
+ 0.8766614754162534,
101
+ 0.8743087945808291
102
+ ],
103
+ [
104
+ 0.8729077459005785,
105
+ 0.8779437978854354,
106
+ 0.8755559808920815
107
+ ],
108
+ [
109
+ 0.873175848841043,
110
+ 0.8781832818585541,
111
+ 0.8757574155692655
112
+ ]
113
+ ],
114
+ "val_fve_aligned_by_layer": [
115
+ [
116
+ 0.8233840016170322,
117
+ 0.825940442334919,
118
+ 0.8251950759538181
119
+ ],
120
+ [
121
+ 0.8476839031224476,
122
+ 0.8521280316782247,
123
+ 0.8511595254793217
124
+ ],
125
+ [
126
+ 0.858345750738813,
127
+ 0.8632998816005847,
128
+ 0.8617524373593755
129
+ ],
130
+ [
131
+ 0.8641392950612213,
132
+ 0.8692198473745616,
133
+ 0.8673227601026365
134
+ ],
135
+ [
136
+ 0.8685973614298236,
137
+ 0.8736037952113526,
138
+ 0.8715948927464909
139
+ ],
140
+ [
141
+ 0.8712089548560338,
142
+ 0.8761506876396259,
143
+ 0.8739966572267223
144
+ ],
145
+ [
146
+ 0.8725146145720757,
147
+ 0.8774263980500985,
148
+ 0.8751902630191823
149
+ ],
150
+ [
151
+ 0.8727765308000655,
152
+ 0.8776650928077897,
153
+ 0.875393784795132
154
+ ]
155
+ ],
156
+ "train_fve_base_by_layer": [
157
+ [
158
+ 0.7016964963285532,
159
+ 0.7011062915187375,
160
+ 0.7000558319247681
161
+ ],
162
+ [
163
+ 0.838823730355456,
164
+ 0.8433957115224558,
165
+ 0.8422748323984475
166
+ ],
167
+ [
168
+ 0.8563592615213856,
169
+ 0.8613082306536749,
170
+ 0.8598773517040781
171
+ ],
172
+ [
173
+ 0.865255449539169,
174
+ 0.8708288593732051,
175
+ 0.8691897918728452
176
+ ],
177
+ [
178
+ 0.8718092898423395,
179
+ 0.8773209839206791,
180
+ 0.8755165421677416
181
+ ],
182
+ [
183
+ 0.876150603852492,
184
+ 0.8818598225042963,
185
+ 0.8800080236606409
186
+ ],
187
+ [
188
+ 0.8793807434047514,
189
+ 0.8849946516198995,
190
+ 0.883135499796948
191
+ ],
192
+ [
193
+ 0.8811041608374507,
194
+ 0.8866949073069985,
195
+ 0.8848176027207227
196
+ ]
197
+ ],
198
+ "train_fve_aligned_by_layer": [
199
+ [
200
+ 0.6981886976376237,
201
+ 0.6973937544917337,
202
+ 0.6972548026685053
203
+ ],
204
+ [
205
+ 0.8377145442848517,
206
+ 0.8420214018345714,
207
+ 0.8410983897410703
208
+ ],
209
+ [
210
+ 0.8557311300290514,
211
+ 0.8604886952855054,
212
+ 0.8592272625982726
213
+ ],
214
+ [
215
+ 0.8647247337487728,
216
+ 0.8699760436315809,
217
+ 0.8686401396554879
218
+ ],
219
+ [
220
+ 0.871257510751712,
221
+ 0.8765467838870804,
222
+ 0.8749307167509183
223
+ ],
224
+ [
225
+ 0.875646241596409,
226
+ 0.8810545213724951,
227
+ 0.8793973550765889
228
+ ],
229
+ [
230
+ 0.8788073496100364,
231
+ 0.8841580943760646,
232
+ 0.8824782781411665
233
+ ],
234
+ [
235
+ 0.8805134411242296,
236
+ 0.8858426592883073,
237
+ 0.8841365650600672
238
+ ]
239
+ ],
240
+ "dead_neurons": [
241
+ 0.9451254373465235,
242
+ 0.9706496352420101,
243
+ 0.9707474327198747,
244
+ 0.9705583059989564,
245
+ 0.9702848919079586,
246
+ 0.9700425463466009,
247
+ 0.9698690966563273,
248
+ 0.9698438135351414
249
+ ],
250
+ "l0_base": [
251
+ 209.83105200630004,
252
+ 188.89718229758134,
253
+ 188.9450576523905,
254
+ 191.0756520607008,
255
+ 192.9378394791068,
256
+ 194.2250183871526,
257
+ 194.89630656957766,
258
+ 195.12645257792832
259
+ ],
260
+ "l0_aligned": [
261
+ 210.18967629703803,
262
+ 189.2792483818023,
263
+ 189.1256618466352,
264
+ 191.12680524925148,
265
+ 192.91545546799463,
266
+ 194.14990057112945,
267
+ 194.79371505888736,
268
+ 195.01710472841813
269
+ ],
270
+ "l0_base_by_layer": [
271
+ [
272
+ 205.6260763280794,
273
+ 209.88612813777,
274
+ 213.98093622300058
275
+ ],
276
+ [
277
+ 182.01882661996498,
278
+ 189.60584500875657,
279
+ 195.066860040864
280
+ ],
281
+ [
282
+ 181.92784953298306,
283
+ 190.21716287215412,
284
+ 194.69014521307648
285
+ ],
286
+ [
287
+ 184.42091725043784,
288
+ 192.66653896672506,
289
+ 196.1394848219498
290
+ ],
291
+ [
292
+ 186.74636967308814,
293
+ 194.6657362813777,
294
+ 197.40139740221832
295
+ ],
296
+ [
297
+ 188.41885580852306,
298
+ 195.99750072971395,
299
+ 198.25868359603035
300
+ ],
301
+ [
302
+ 189.31253648569762,
303
+ 196.66821730881495,
304
+ 198.7081509048453
305
+ ],
306
+ [
307
+ 189.62868505545825,
308
+ 196.90103254524226,
309
+ 198.84962419731465
310
+ ]
311
+ ],
312
+ "l0_aligned_by_layer": [
313
+ [
314
+ 206.34398715703443,
315
+ 209.85872737886748,
316
+ 214.36629816112085
317
+ ],
318
+ [
319
+ 182.57375583771162,
320
+ 189.64152802101577,
321
+ 195.62244600116753
322
+ ],
323
+ [
324
+ 182.25118578517223,
325
+ 190.08411777583188,
326
+ 195.04166666666666
327
+ ],
328
+ [
329
+ 184.60119308231174,
330
+ 192.41106611208406,
331
+ 196.36814068884996
332
+ ],
333
+ [
334
+ 186.8388244308231,
335
+ 194.3693812025686,
336
+ 197.53814579684763
337
+ ],
338
+ [
339
+ 188.42837857559837,
340
+ 195.65619527145358,
341
+ 198.36511237594863
342
+ ],
343
+ [
344
+ 189.29121059544659,
345
+ 196.3086142732049,
346
+ 198.78130472854642
347
+ ],
348
+ [
349
+ 189.5982742265032,
350
+ 196.52698117338002,
351
+ 198.92604349095154
352
+ ]
353
+ ],
354
+ "val_l0_base": [
355
+ 191.54323541930833,
356
+ 189.07813953978854,
357
+ 190.84690026088535,
358
+ 192.91267203286057,
359
+ 194.8115054624867,
360
+ 195.63816665729303,
361
+ 196.18994100680527,
362
+ 196.35193550149808
363
+ ],
364
+ "val_l0_aligned": [
365
+ 192.0186752399225,
366
+ 189.2518055701131,
367
+ 190.91334501735827,
368
+ 192.87349672966602,
369
+ 194.75840359832603,
370
+ 195.5591781376544,
371
+ 196.06355988667272,
372
+ 196.22006401341622
373
+ ],
374
+ "val_l0_base_by_layer": [
375
+ [
376
+ 185.64757853403142,
377
+ 191.61927356020942,
378
+ 197.36283818589456
379
+ ],
380
+ [
381
+ 182.28844895287958,
382
+ 189.9172666060363,
383
+ 195.02868676310433
384
+ ],
385
+ [
386
+ 183.83273347385267,
387
+ 192.34304106547572,
388
+ 196.36491058509387
389
+ ],
390
+ [
391
+ 186.74291017542336,
392
+ 194.58295157068062,
393
+ 197.41214005235602
394
+ ],
395
+ [
396
+ 188.8255345209731,
397
+ 196.6678664921466,
398
+ 198.9410994764398
399
+ ],
400
+ [
401
+ 190.03910340314135,
402
+ 197.35323952879583,
403
+ 199.52214226048653
404
+ ],
405
+ [
406
+ 190.72627617801047,
407
+ 197.96722300264847,
408
+ 199.87630890052355
409
+ ],
410
+ [
411
+ 190.95555194635043,
412
+ 198.10874787175842,
413
+ 199.99149214659687
414
+ ]
415
+ ],
416
+ "val_l0_aligned_by_layer": [
417
+ [
418
+ 186.1591405419155,
419
+ 191.84249566362791,
420
+ 198.05437393587923
421
+ ],
422
+ [
423
+ 182.5016906898059,
424
+ 189.92092064662754,
425
+ 195.33278795811518
426
+ ],
427
+ [
428
+ 184.04439356938707,
429
+ 192.4237020602401,
430
+ 196.27192408376965
431
+ ],
432
+ [
433
+ 186.73592931937173,
434
+ 194.31893545420382,
435
+ 197.56560863874344
436
+ ],
437
+ [
438
+ 188.91552143196785,
439
+ 196.25267236519858,
440
+ 199.10700261780104
441
+ ],
442
+ [
443
+ 190.11970991364325,
444
+ 196.97284031413614,
445
+ 199.58496948561742
446
+ ],
447
+ [
448
+ 190.69617146596858,
449
+ 197.53015927619336,
450
+ 199.96433246073298
451
+ ],
452
+ [
453
+ 190.91961174610398,
454
+ 197.67152053273784,
455
+ 200.0690445026178
456
+ ]
457
+ ],
458
+ "self_recon": [
459
+ 3.147098605492208,
460
+ 1.8534241820453414,
461
+ 1.2700124087037221,
462
+ 0.978502946956354,
463
+ 0.7723350193649443,
464
+ 0.6721525268114873,
465
+ 0.5956131902899021,
466
+ 0.5710704998704195
467
+ ],
468
+ "cross_recon": [
469
+ 3.1112218126176185,
470
+ 1.8204483565919523,
471
+ 1.26578516806696,
472
+ 0.9781602442751693,
473
+ 0.7761116657978042,
474
+ 0.6765880859601031,
475
+ 0.6027559136794592,
476
+ 0.5784055959809543
477
+ ],
478
+ "sparsity": [
479
+ 0.21780632954293083,
480
+ 0.1109368161880615,
481
+ 0.11740272650054791,
482
+ 0.12095141849250593,
483
+ 0.12291998712381841,
484
+ 0.12457959489191198,
485
+ 0.1256028851021192,
486
+ 0.12605826421415187
487
+ ],
488
+ "val_self_recon": [
489
+ 2.677790811236616,
490
+ 1.942986978286224,
491
+ 1.110527089597043,
492
+ 1.0172813413030815,
493
+ 0.8395302370580703,
494
+ 0.709914362711432,
495
+ 0.6143848535590147,
496
+ 0.6117264227405268
497
+ ],
498
+ "val_cross_recon": [
499
+ 2.640897078663891,
500
+ 1.943937375127333,
501
+ 1.1101361076245133,
502
+ 1.0132286010300302,
503
+ 0.8250547973585378,
504
+ 0.7056703068198958,
505
+ 0.6165956046568786,
506
+ 0.6138372463393585
507
+ ],
508
+ "val_sparsity": [
509
+ 0.10814674432677124,
510
+ 0.11746669529464232,
511
+ 0.12130338378757706,
512
+ 0.12336198517477324,
513
+ 0.12571066255188737,
514
+ 0.12654356826194293,
515
+ 0.12717439632141153,
516
+ 0.12746002025785247
517
+ ],
518
+ "layers": [
519
+ 21,
520
+ 22,
521
+ 23
522
+ ],
523
+ "topk_mode": "model_balanced_layer_agg"
524
+ }
v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/aligned_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: e7029e11b2a0baac5db015215998f1e355ada9f9a96225079b72e543dc9e59c4
  • Pointer size: 131 Bytes
  • Size of remote file: 181 kB
v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/base_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: ceb0bc615562dc77fd6ec303cf5b5ee636e1e28f09259dcb4ce353271930ed22
  • Pointer size: 131 Bytes
  • Size of remote file: 180 kB
v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/cf_shift_by_layer.png ADDED

Git LFS Details

  • SHA256: 60fc72cf8031671b7ade8472fce73e6c97358b21f849674855f41929220ed394
  • Pointer size: 130 Bytes
  • Size of remote file: 44.3 kB
v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/cf_shift_p95_by_layer.png ADDED

Git LFS Details

  • SHA256: 31b5956c95ad283a9fdb84d7455c0ab6a06c3050507f0223455aea6bdbd42d24
  • Pointer size: 131 Bytes
  • Size of remote file: 110 kB
v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/class_distribution_multilayer.png ADDED

Git LFS Details

  • SHA256: d7562609167b801533edec58b0bd4bad47852dd7370ebf1d17001795efb28d2f
  • Pointer size: 130 Bytes
  • Size of remote file: 93.7 kB
v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/class_distribution_primary.png ADDED

Git LFS Details

  • SHA256: 6811b68413db92de41059e68c6c5b4025d1612950d3e88d6311492a65e5be29d
  • Pointer size: 131 Bytes
  • Size of remote file: 145 kB
v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/cross_layer_cosine_drift_by_stream.png ADDED

Git LFS Details

  • SHA256: a1f0e9038e380a5aa7414fe9559641bbee59ff5cefe465b3a5c4a21a9a8077bd
  • Pointer size: 130 Bytes
  • Size of remote file: 44.6 kB
v1-undetermined/assembled_activations/qwen3-4b-dpo/plots/decoder_norm_ratio_by_layer.png ADDED

Git LFS Details

  • SHA256: ddd676f02c78e399effa493fe9fd61af604abe07e5b5f9afd7c49126eeb11835
  • Pointer size: 130 Bytes
  • Size of remote file: 47.6 kB