ishangarg183 commited on
Commit
b7a396a
·
verified ·
1 Parent(s): 92fa9b2

Rename v2-samelayer -> v1-undetermined (1/14)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +11 -0
  2. v1-undetermined/assembled_activations/llama32-3b-dpo/checkpoints/final.pt +3 -0
  3. v1-undetermined/assembled_activations/llama32-3b-dpo/features/counterfactual_scores.csv +0 -0
  4. v1-undetermined/assembled_activations/llama32-3b-dpo/features/counterfactual_scores_by_layer.csv +0 -0
  5. v1-undetermined/assembled_activations/llama32-3b-dpo/features/cross_layer_cosine_drift.csv +3 -0
  6. v1-undetermined/assembled_activations/llama32-3b-dpo/features/decoder_layer_profiles.csv +3 -0
  7. v1-undetermined/assembled_activations/llama32-3b-dpo/features/feature_classification.csv +0 -0
  8. v1-undetermined/assembled_activations/llama32-3b-dpo/features/merged_classification.csv +0 -0
  9. v1-undetermined/assembled_activations/llama32-3b-dpo/features/model_layer_stream_patterns.csv +0 -0
  10. v1-undetermined/assembled_activations/llama32-3b-dpo/features/superposition_analysis.json +0 -0
  11. v1-undetermined/assembled_activations/llama32-3b-dpo/metrics/aggregate_metrics.json +388 -0
  12. v1-undetermined/assembled_activations/llama32-3b-dpo/metrics/training_metrics.json +524 -0
  13. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/aligned_decoder_norm_heatmap.png +3 -0
  14. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/base_decoder_norm_heatmap.png +3 -0
  15. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/cf_shift_by_layer.png +3 -0
  16. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/cf_shift_p95_by_layer.png +3 -0
  17. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/class_distribution_multilayer.png +3 -0
  18. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/class_distribution_primary.png +3 -0
  19. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/cross_layer_cosine_drift_by_stream.png +3 -0
  20. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/decoder_norm_ratio_by_layer.png +3 -0
  21. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/feature_layer_trajectories.png +3 -0
  22. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/feature_sharing_ratio_by_layer.png +3 -0
  23. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/fve_by_layer.png +3 -0
  24. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/l0_by_layer.png +3 -0
  25. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/layer_concentration_entropy.png +3 -0
  26. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/loss_curves.png +3 -0
  27. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/max_norm_layer_migration.png +3 -0
  28. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/rho_histogram_by_layer.png +3 -0
  29. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/rho_theta_scatter_by_layer.png +3 -0
  30. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/superposition_by_layer.png +3 -0
  31. v1-undetermined/assembled_activations/llama32-3b-dpo/plots/theta_by_layer.png +3 -0
  32. v1-undetermined/assembled_activations/llama32-3b-grpo/checkpoints/final.pt +3 -0
  33. v1-undetermined/assembled_activations/llama32-3b-grpo/features/counterfactual_scores.csv +0 -0
  34. v1-undetermined/assembled_activations/llama32-3b-grpo/features/counterfactual_scores_by_layer.csv +0 -0
  35. v1-undetermined/assembled_activations/llama32-3b-grpo/features/cross_layer_cosine_drift.csv +3 -0
  36. v1-undetermined/assembled_activations/llama32-3b-grpo/features/decoder_layer_profiles.csv +3 -0
  37. v1-undetermined/assembled_activations/llama32-3b-grpo/features/feature_classification.csv +0 -0
  38. v1-undetermined/assembled_activations/llama32-3b-grpo/features/merged_classification.csv +0 -0
  39. v1-undetermined/assembled_activations/llama32-3b-grpo/features/model_layer_stream_patterns.csv +0 -0
  40. v1-undetermined/assembled_activations/llama32-3b-grpo/features/superposition_analysis.json +0 -0
  41. v1-undetermined/assembled_activations/llama32-3b-grpo/metrics/aggregate_metrics.json +388 -0
  42. v1-undetermined/assembled_activations/llama32-3b-grpo/metrics/training_metrics.json +524 -0
  43. v1-undetermined/assembled_activations/llama32-3b-grpo/plots/aligned_decoder_norm_heatmap.png +3 -0
  44. v1-undetermined/assembled_activations/llama32-3b-grpo/plots/base_decoder_norm_heatmap.png +3 -0
  45. v1-undetermined/assembled_activations/llama32-3b-grpo/plots/cf_shift_by_layer.png +3 -0
  46. v1-undetermined/assembled_activations/llama32-3b-grpo/plots/cf_shift_p95_by_layer.png +3 -0
  47. v1-undetermined/assembled_activations/llama32-3b-grpo/plots/class_distribution_multilayer.png +3 -0
  48. v1-undetermined/assembled_activations/llama32-3b-grpo/plots/class_distribution_primary.png +3 -0
  49. v1-undetermined/assembled_activations/llama32-3b-grpo/plots/cross_layer_cosine_drift_by_stream.png +3 -0
  50. v1-undetermined/assembled_activations/llama32-3b-grpo/plots/decoder_norm_ratio_by_layer.png +3 -0
.gitattributes CHANGED
@@ -313,3 +313,14 @@ v3-samelayer/samelayer_separate/qwen3-4b-ppo/singlelayer_separate/L22/features/s
313
  v3-samelayer/samelayer_separate/qwen3-4b-simpo/singlelayer_separate/L20/features/superposition_analysis.json filter=lfs diff=lfs merge=lfs -text
314
  v3-samelayer/samelayer_separate/qwen3-4b-simpo/singlelayer_separate/L21/features/superposition_analysis.json filter=lfs diff=lfs merge=lfs -text
315
  v3-samelayer/samelayer_separate/qwen3-4b-simpo/singlelayer_separate/L22/features/superposition_analysis.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
313
  v3-samelayer/samelayer_separate/qwen3-4b-simpo/singlelayer_separate/L20/features/superposition_analysis.json filter=lfs diff=lfs merge=lfs -text
314
  v3-samelayer/samelayer_separate/qwen3-4b-simpo/singlelayer_separate/L21/features/superposition_analysis.json filter=lfs diff=lfs merge=lfs -text
315
  v3-samelayer/samelayer_separate/qwen3-4b-simpo/singlelayer_separate/L22/features/superposition_analysis.json filter=lfs diff=lfs merge=lfs -text
316
+ v1-undetermined/assembled_activations/llama32-3b-dpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
317
+ v1-undetermined/assembled_activations/llama32-3b-dpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
318
+ v1-undetermined/assembled_activations/llama32-3b-grpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
319
+ v1-undetermined/assembled_activations/llama32-3b-grpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
320
+ v1-undetermined/assembled_activations/llama32-3b-kto/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
321
+ v1-undetermined/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
322
+ v1-undetermined/assembled_activations/llama32-3b-orpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
323
+ v1-undetermined/assembled_activations/llama32-3b-orpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
324
+ v1-undetermined/assembled_activations/llama32-3b-ppo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
325
+ v1-undetermined/assembled_activations/llama32-3b-ppo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
326
+ v1-undetermined/assembled_activations/llama32-3b-ppo/features/model_layer_stream_patterns.csv filter=lfs diff=lfs merge=lfs -text
v1-undetermined/assembled_activations/llama32-3b-dpo/checkpoints/final.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00fcd03d23e426a857fd9d36297fc24efae3c9d071c11a0fc88553105cce0ae4
3
+ size 10873439922
v1-undetermined/assembled_activations/llama32-3b-dpo/features/counterfactual_scores.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-dpo/features/counterfactual_scores_by_layer.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-dpo/features/cross_layer_cosine_drift.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98369ea77bd0d42503293df68eed91eedeb05452b6d12ffda8e9f6d552ee09e0
3
+ size 27252306
v1-undetermined/assembled_activations/llama32-3b-dpo/features/decoder_layer_profiles.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f685534b39b1dcd0f85a3934158b2fb28e947472c0c6b9169aa990ac896fa897
3
+ size 13008950
v1-undetermined/assembled_activations/llama32-3b-dpo/features/feature_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-dpo/features/merged_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-dpo/features/model_layer_stream_patterns.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-dpo/features/superposition_analysis.json ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-dpo/metrics/aggregate_metrics.json ADDED
@@ -0,0 +1,388 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crosscoder_kind": "multilayer_sparc",
3
+ "layers": [
4
+ 12,
5
+ 13,
6
+ 14
7
+ ],
8
+ "topk_mode": "model_balanced_layer_agg",
9
+ "topk": 400,
10
+ "expansion_factor": 8,
11
+ "dict_size": 24576,
12
+ "forced_shared_fraction": 0.06,
13
+ "class_counts": {
14
+ "aligned_only": 6255,
15
+ "shared_redirected": 5381,
16
+ "shared_intermediate": 5281,
17
+ "base_only": 3273,
18
+ "shared_attenuated": 2898,
19
+ "shared_aligned": 1488
20
+ },
21
+ "multilayer_class_counts": {
22
+ "drifting_or_rotating": 12229,
23
+ "persistent_aligned_only": 4659,
24
+ "persistent_shared": 2467,
25
+ "persistent_base_only": 2016,
26
+ "localized_aligned_only": 1596,
27
+ "localized_base_only": 1257,
28
+ "mixed_or_ambiguous": 352
29
+ },
30
+ "classification_thresholds": {
31
+ "rho_base_only": 0.4,
32
+ "rho_aligned_only": 0.5571486335286551,
33
+ "rho_shared_low": 0.441999366421853,
34
+ "rho_shared_high": 0.5571486335286551
35
+ },
36
+ "threshold_sensitivity": {
37
+ "original": {
38
+ "aligned_only": 6255,
39
+ "shared_redirected": 5381,
40
+ "shared_intermediate": 5281,
41
+ "base_only": 3273,
42
+ "shared_attenuated": 2898,
43
+ "shared_aligned": 1488
44
+ },
45
+ "perturbed": {
46
+ "delta_-0.05": {
47
+ "base_only": 840,
48
+ "aligned_only": 2834,
49
+ "shared_aligned": 1500,
50
+ "shared_redirected": 9107,
51
+ "shared_intermediate": 8349,
52
+ "shared_attenuated": 1946,
53
+ "other": 0
54
+ },
55
+ "delta_+0.05": {
56
+ "base_only": 6820,
57
+ "aligned_only": 10851,
58
+ "shared_aligned": 1476,
59
+ "shared_redirected": 734,
60
+ "shared_intermediate": 804,
61
+ "shared_attenuated": 3891,
62
+ "other": 0
63
+ }
64
+ },
65
+ "perturbation": 0.05
66
+ },
67
+ "class_counts_by_layer": {
68
+ "12": {
69
+ "shared_redirected": 7895,
70
+ "shared_intermediate": 6656,
71
+ "base_only": 4688,
72
+ "shared_aligned": 2938,
73
+ "aligned_only": 2399
74
+ },
75
+ "13": {
76
+ "shared_redirected": 8816,
77
+ "shared_intermediate": 8466,
78
+ "base_only": 3035,
79
+ "shared_aligned": 2802,
80
+ "aligned_only": 1457
81
+ },
82
+ "14": {
83
+ "shared_redirected": 8569,
84
+ "base_only": 4460,
85
+ "shared_intermediate": 4299,
86
+ "aligned_only": 3222,
87
+ "shared_attenuated": 2072,
88
+ "shared_aligned": 1954
89
+ }
90
+ },
91
+ "feature_sharing_ratio_by_layer": {
92
+ "12": 0.7116292317708334,
93
+ "13": 0.8172200520833334,
94
+ "14": 0.6874186197916666
95
+ },
96
+ "decoder_amplification_by_layer": {
97
+ "12": {
98
+ "median": 0.9999999872344182,
99
+ "p95": 4.808803178398062
100
+ },
101
+ "13": {
102
+ "median": 0.9999999877864876,
103
+ "p95": 2.875206066511721
104
+ },
105
+ "14": {
106
+ "median": 0.9999999891683586,
107
+ "p95": 2.0530812721600027
108
+ }
109
+ },
110
+ "classification_thresholds_by_layer": {
111
+ "12": {
112
+ "rho_base_only": 0.3249751650747221,
113
+ "rho_aligned_only": 0.7504738322947574,
114
+ "rho_shared_low": 0.3249751650747221,
115
+ "rho_shared_high": 0.7504738322947574
116
+ },
117
+ "13": {
118
+ "rho_base_only": 0.33311266648506177,
119
+ "rho_aligned_only": 0.7273791891804369,
120
+ "rho_shared_low": 0.33311266648506177,
121
+ "rho_shared_high": 0.7273791891804369
122
+ },
123
+ "14": {
124
+ "rho_base_only": 0.4,
125
+ "rho_aligned_only": 0.6195673090879186,
126
+ "rho_shared_low": 0.4332179366876391,
127
+ "rho_shared_high": 0.6195673090879186
128
+ }
129
+ },
130
+ "threshold_sensitivity_by_layer": {
131
+ "12": {
132
+ "original": {
133
+ "shared_redirected": 7895,
134
+ "shared_intermediate": 6656,
135
+ "base_only": 4688,
136
+ "shared_aligned": 2938,
137
+ "aligned_only": 2399
138
+ },
139
+ "perturbed": {
140
+ "delta_-0.05": {
141
+ "base_only": 3192,
142
+ "aligned_only": 1558,
143
+ "shared_aligned": 3589,
144
+ "shared_redirected": 8800,
145
+ "shared_intermediate": 7437,
146
+ "shared_attenuated": 0,
147
+ "other": 0
148
+ },
149
+ "delta_+0.05": {
150
+ "base_only": 6578,
151
+ "aligned_only": 3609,
152
+ "shared_aligned": 2147,
153
+ "shared_redirected": 6691,
154
+ "shared_intermediate": 5551,
155
+ "shared_attenuated": 0,
156
+ "other": 0
157
+ }
158
+ },
159
+ "perturbation": 0.05
160
+ },
161
+ "13": {
162
+ "original": {
163
+ "shared_redirected": 8816,
164
+ "shared_intermediate": 8466,
165
+ "base_only": 3035,
166
+ "shared_aligned": 2802,
167
+ "aligned_only": 1457
168
+ },
169
+ "perturbed": {
170
+ "delta_-0.05": {
171
+ "base_only": 1793,
172
+ "aligned_only": 829,
173
+ "shared_aligned": 3190,
174
+ "shared_redirected": 9548,
175
+ "shared_intermediate": 9216,
176
+ "shared_attenuated": 0,
177
+ "other": 0
178
+ },
179
+ "delta_+0.05": {
180
+ "base_only": 5100,
181
+ "aligned_only": 2534,
182
+ "shared_aligned": 2092,
183
+ "shared_redirected": 7747,
184
+ "shared_intermediate": 7103,
185
+ "shared_attenuated": 0,
186
+ "other": 0
187
+ }
188
+ },
189
+ "perturbation": 0.05
190
+ },
191
+ "14": {
192
+ "original": {
193
+ "shared_redirected": 8569,
194
+ "base_only": 4460,
195
+ "shared_intermediate": 4299,
196
+ "aligned_only": 3222,
197
+ "shared_attenuated": 2072,
198
+ "shared_aligned": 1954
199
+ },
200
+ "perturbed": {
201
+ "delta_-0.05": {
202
+ "base_only": 1933,
203
+ "aligned_only": 1278,
204
+ "shared_aligned": 3019,
205
+ "shared_redirected": 10753,
206
+ "shared_intermediate": 6024,
207
+ "shared_attenuated": 1569,
208
+ "other": 0
209
+ },
210
+ "delta_+0.05": {
211
+ "base_only": 7807,
212
+ "aligned_only": 6199,
213
+ "shared_aligned": 1340,
214
+ "shared_redirected": 4592,
215
+ "shared_intermediate": 2047,
216
+ "shared_attenuated": 2591,
217
+ "other": 0
218
+ }
219
+ },
220
+ "perturbation": 0.05
221
+ }
222
+ },
223
+ "counterfactual_shift_by_layer": {
224
+ "12": {
225
+ "aligned_only": {
226
+ "mean_shift": 2.4112238498396065e-06,
227
+ "median_shift": 0.0,
228
+ "p95_abs_shift": 0.0,
229
+ "count": 6255
230
+ },
231
+ "base_only": {
232
+ "mean_shift": -2.4341085855552528e-06,
233
+ "median_shift": 0.0,
234
+ "p95_abs_shift": 0.0,
235
+ "count": 3273
236
+ },
237
+ "shared_aligned": {
238
+ "mean_shift": -7.11400898647972e-05,
239
+ "median_shift": 0.0,
240
+ "p95_abs_shift": 0.08343747109174728,
241
+ "count": 1488
242
+ },
243
+ "shared_attenuated": {
244
+ "mean_shift": -1.0171309737362479e-07,
245
+ "median_shift": 0.0,
246
+ "p95_abs_shift": 0.0,
247
+ "count": 2898
248
+ },
249
+ "shared_intermediate": {
250
+ "mean_shift": 5.1859217087668957e-08,
251
+ "median_shift": 0.0,
252
+ "p95_abs_shift": 0.0,
253
+ "count": 5281
254
+ },
255
+ "shared_redirected": {
256
+ "mean_shift": 2.86653050904218e-09,
257
+ "median_shift": 0.0,
258
+ "p95_abs_shift": 0.0,
259
+ "count": 5381
260
+ }
261
+ },
262
+ "13": {
263
+ "aligned_only": {
264
+ "mean_shift": 8.161741125300389e-06,
265
+ "median_shift": 0.0,
266
+ "p95_abs_shift": 0.0,
267
+ "count": 6255
268
+ },
269
+ "base_only": {
270
+ "mean_shift": -1.1108721304735658e-05,
271
+ "median_shift": 0.0,
272
+ "p95_abs_shift": 0.0,
273
+ "count": 3273
274
+ },
275
+ "shared_aligned": {
276
+ "mean_shift": 5.4217558803460264e-05,
277
+ "median_shift": 0.0,
278
+ "p95_abs_shift": 0.09117132201790806,
279
+ "count": 1488
280
+ },
281
+ "shared_attenuated": {
282
+ "mean_shift": -2.378168234298194e-06,
283
+ "median_shift": 0.0,
284
+ "p95_abs_shift": 0.0,
285
+ "count": 2898
286
+ },
287
+ "shared_intermediate": {
288
+ "mean_shift": -4.839519160503355e-08,
289
+ "median_shift": 0.0,
290
+ "p95_abs_shift": 0.0,
291
+ "count": 5281
292
+ },
293
+ "shared_redirected": {
294
+ "mean_shift": 1.3520313741632717e-07,
295
+ "median_shift": 0.0,
296
+ "p95_abs_shift": 0.0,
297
+ "count": 5381
298
+ }
299
+ },
300
+ "14": {
301
+ "aligned_only": {
302
+ "mean_shift": 3.1487233612113166e-05,
303
+ "median_shift": 0.0,
304
+ "p95_abs_shift": 0.0,
305
+ "count": 6255
306
+ },
307
+ "base_only": {
308
+ "mean_shift": -3.585443771814724e-05,
309
+ "median_shift": 0.0,
310
+ "p95_abs_shift": 0.0,
311
+ "count": 3273
312
+ },
313
+ "shared_aligned": {
314
+ "mean_shift": 6.549898494149471e-05,
315
+ "median_shift": 3.887436008653822e-07,
316
+ "p95_abs_shift": 0.10445111840963363,
317
+ "count": 1488
318
+ },
319
+ "shared_attenuated": {
320
+ "mean_shift": -1.079474320873924e-05,
321
+ "median_shift": 0.0,
322
+ "p95_abs_shift": 0.0,
323
+ "count": 2898
324
+ },
325
+ "shared_intermediate": {
326
+ "mean_shift": 8.715442941091603e-08,
327
+ "median_shift": 0.0,
328
+ "p95_abs_shift": 0.0,
329
+ "count": 5281
330
+ },
331
+ "shared_redirected": {
332
+ "mean_shift": 8.519746033905039e-07,
333
+ "median_shift": 0.0,
334
+ "p95_abs_shift": 0.0,
335
+ "count": 5381
336
+ }
337
+ }
338
+ },
339
+ "total_features": 24576,
340
+ "fve_base": 0.6943563584881928,
341
+ "fve_aligned": 0.6857018002664855,
342
+ "fve_base_by_layer": [
343
+ 0.6758359645673742,
344
+ 0.6896311734359302,
345
+ 0.7176018759842319
346
+ ],
347
+ "fve_aligned_by_layer": [
348
+ 0.6675708861875285,
349
+ 0.6810908585942853,
350
+ 0.7084435948526672
351
+ ],
352
+ "val_fve_base_by_layer": [
353
+ 0.6758359645673742,
354
+ 0.6896311734359302,
355
+ 0.7176018759842319
356
+ ],
357
+ "val_fve_aligned_by_layer": [
358
+ 0.6675708861875285,
359
+ 0.6810908585942853,
360
+ 0.7084435948526672
361
+ ],
362
+ "dead_neuron_fraction": 0.9607996527590532,
363
+ "l0_sparsity_base": 75.38008013901764,
364
+ "l0_sparsity_aligned": 74.12330600273101,
365
+ "l0_base_by_layer": [
366
+ 63.096559398715705,
367
+ 75.58358873321657,
368
+ 87.46008464681844
369
+ ],
370
+ "l0_aligned_by_layer": [
371
+ 61.64835084646818,
372
+ 74.61715557501459,
373
+ 86.1044038237011
374
+ ],
375
+ "val_l0_base": 76.88609194131422,
376
+ "val_l0_aligned": 75.5361974426589,
377
+ "val_l0_base_by_layer": [
378
+ 64.24105585307976,
379
+ 77.04439356938707,
380
+ 89.37281851244222
381
+ ],
382
+ "val_l0_aligned_by_layer": [
383
+ 62.69873474280872,
384
+ 75.9856566583923,
385
+ 87.92419285799197
386
+ ],
387
+ "superposition_fraction": 0.0
388
+ }
v1-undetermined/assembled_activations/llama32-3b-dpo/metrics/training_metrics.json ADDED
@@ -0,0 +1,524 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epochs": [
3
+ 1,
4
+ 2,
5
+ 3,
6
+ 4,
7
+ 5,
8
+ 6,
9
+ 7,
10
+ 8
11
+ ],
12
+ "train_loss": [
13
+ 0.07673105311085798,
14
+ 0.04314152776142414,
15
+ 0.04106981019553078,
16
+ 0.039948021885392544,
17
+ 0.039125458598876305,
18
+ 0.038466410324308024,
19
+ 0.03799404853697207,
20
+ 0.03772751730255091
21
+ ],
22
+ "val_loss": [
23
+ 0.0452033090380786,
24
+ 0.04240399528860422,
25
+ 0.04129640658566465,
26
+ 0.04064904576586803,
27
+ 0.0401821202008512,
28
+ 0.03991277757739521,
29
+ 0.03979530702320693,
30
+ 0.03977678012083338
31
+ ],
32
+ "train_fve_base": [
33
+ 0.5215789154168545,
34
+ 0.6578964054480116,
35
+ 0.6789237529787485,
36
+ 0.6908673060545779,
37
+ 0.6998354073720443,
38
+ 0.7069360464550359,
39
+ 0.7119182502047332,
40
+ 0.7145967230905376
41
+ ],
42
+ "train_fve_aligned": [
43
+ 0.5162195712566496,
44
+ 0.6509691739179764,
45
+ 0.6700791642809221,
46
+ 0.6808612814996233,
47
+ 0.6889950146956257,
48
+ 0.6955827845165901,
49
+ 0.7002096909288071,
50
+ 0.7027074147300698
51
+ ],
52
+ "val_fve_base": [
53
+ 0.6388902236653872,
54
+ 0.6665667504540289,
55
+ 0.6772977547495777,
56
+ 0.6839563450264057,
57
+ 0.6892214420578242,
58
+ 0.6923850852781566,
59
+ 0.6940219399192571,
60
+ 0.6943563584881928
61
+ ],
62
+ "val_fve_aligned": [
63
+ 0.6334137236260619,
64
+ 0.6591302406101327,
65
+ 0.6693885479297937,
66
+ 0.6759788746609113,
67
+ 0.6809322253571755,
68
+ 0.6837883154759232,
69
+ 0.6854131986333437,
70
+ 0.6857018002664855
71
+ ],
72
+ "val_fve_base_by_layer": [
73
+ [
74
+ 0.6264502502860824,
75
+ 0.633638423774879,
76
+ 0.6565819432598134
77
+ ],
78
+ [
79
+ 0.6513919614996585,
80
+ 0.6613985574682345,
81
+ 0.6869096593707019
82
+ ],
83
+ [
84
+ 0.6611033381591916,
85
+ 0.6716362203603016,
86
+ 0.6991536592313756
87
+ ],
88
+ [
89
+ 0.66649499051858,
90
+ 0.6792081759862251,
91
+ 0.7061658099059659
92
+ ],
93
+ [
94
+ 0.6711723985472274,
95
+ 0.6845272938618485,
96
+ 0.7119645825855395
97
+ ],
98
+ [
99
+ 0.6738163388836447,
100
+ 0.6877937959750909,
101
+ 0.7155450829036573
102
+ ],
103
+ [
104
+ 0.6754764951960579,
105
+ 0.6893106998573423,
106
+ 0.7172785657238586
107
+ ],
108
+ [
109
+ 0.6758359645673742,
110
+ 0.6896311734359302,
111
+ 0.7176018759842319
112
+ ]
113
+ ],
114
+ "val_fve_aligned_by_layer": [
115
+ [
116
+ 0.6210345457361631,
117
+ 0.6278264157435033,
118
+ 0.651380156659331
119
+ ],
120
+ [
121
+ 0.6446888418721903,
122
+ 0.6542470848373093,
123
+ 0.6784547342679887
124
+ ],
125
+ [
126
+ 0.6536089899652291,
127
+ 0.6642535906187527,
128
+ 0.6903030060972842
129
+ ],
130
+ [
131
+ 0.6592154839900152,
132
+ 0.6713800143196945,
133
+ 0.6973410641959824
134
+ ],
135
+ [
136
+ 0.6634331439178027,
137
+ 0.6763586255268277,
138
+ 0.7030048576325022
139
+ ],
140
+ [
141
+ 0.6656399893511028,
142
+ 0.6792695619048873,
143
+ 0.7064553421205251
144
+ ],
145
+ [
146
+ 0.6672888201568763,
147
+ 0.6807850318429358,
148
+ 0.7081656861679716
149
+ ],
150
+ [
151
+ 0.6675708861875285,
152
+ 0.6810908585942853,
153
+ 0.7084435948526672
154
+ ]
155
+ ],
156
+ "train_fve_base_by_layer": [
157
+ [
158
+ 0.5129442294299498,
159
+ 0.5156938191818621,
160
+ 0.5360986467905373
161
+ ],
162
+ [
163
+ 0.6431764893133739,
164
+ 0.6530787017494353,
165
+ 0.67743396191867
166
+ ],
167
+ [
168
+ 0.6618652313129983,
169
+ 0.6743674225718114,
170
+ 0.700538551466408
171
+ ],
172
+ [
173
+ 0.6724326563939252,
174
+ 0.6864327243107908,
175
+ 0.7137364810555564
176
+ ],
177
+ [
178
+ 0.6801178986331697,
179
+ 0.6953849959846693,
180
+ 0.7240032671281489
181
+ ],
182
+ [
183
+ 0.6860084016978636,
184
+ 0.7023456112307133,
185
+ 0.7324540658576127
186
+ ],
187
+ [
188
+ 0.6901865686092611,
189
+ 0.707262370611448,
190
+ 0.7383057537025969
191
+ ],
192
+ [
193
+ 0.6925097409564016,
194
+ 0.7099300534907518,
195
+ 0.7413503145587006
196
+ ]
197
+ ],
198
+ "train_fve_aligned_by_layer": [
199
+ [
200
+ 0.5080299372222086,
201
+ 0.5100825478372278,
202
+ 0.5305461785073734
203
+ ],
204
+ [
205
+ 0.6369982385802394,
206
+ 0.6462390441262645,
207
+ 0.6696701796167578
208
+ ],
209
+ [
210
+ 0.6540195649295123,
211
+ 0.6656472806476809,
212
+ 0.6905705887743833
213
+ ],
214
+ [
215
+ 0.6633175232573274,
216
+ 0.6766819278886286,
217
+ 0.7025843324608366
218
+ ],
219
+ [
220
+ 0.6700740303133081,
221
+ 0.6847399561222853,
222
+ 0.7121709983945936
223
+ ],
224
+ [
225
+ 0.6753198687472124,
226
+ 0.6911622862055924,
227
+ 0.7202661430981865
228
+ ],
229
+ [
230
+ 0.679051866255792,
231
+ 0.6957064838139411,
232
+ 0.725870660641565
233
+ ],
234
+ [
235
+ 0.6811704615786698,
236
+ 0.6981789980168518,
237
+ 0.7287727255467706
238
+ ]
239
+ ],
240
+ "dead_neurons": [
241
+ 0.8832501577690757,
242
+ 0.9564017257144837,
243
+ 0.9614509329834808,
244
+ 0.9621241365130436,
245
+ 0.961682792024451,
246
+ 0.961177955297302,
247
+ 0.9608592282452224,
248
+ 0.9607996527590532
249
+ ],
250
+ "l0_base": [
251
+ 120.13765799393796,
252
+ 71.49038236432095,
253
+ 70.5635059042695,
254
+ 71.62221756471365,
255
+ 73.03186058316116,
256
+ 74.30840404842307,
257
+ 75.1019739261233,
258
+ 75.38008013901764
259
+ ],
260
+ "l0_aligned": [
261
+ 118.25455218313455,
262
+ 69.83914911879202,
263
+ 69.18486336275491,
264
+ 70.40499982149088,
265
+ 71.87874234272455,
266
+ 73.09848961771682,
267
+ 73.85036860442064,
268
+ 74.12330600273101
269
+ ],
270
+ "l0_base_by_layer": [
271
+ [
272
+ 110.42726576182136,
273
+ 118.15745402802102,
274
+ 131.82824357851723
275
+ ],
276
+ [
277
+ 60.61651707530648,
278
+ 70.74379743140689,
279
+ 83.11082530647985
280
+ ],
281
+ [
282
+ 59.78093987157035,
283
+ 70.1333917104495,
284
+ 81.77617848803268
285
+ ],
286
+ [
287
+ 60.539860624635146,
288
+ 71.36715557501459,
289
+ 82.95962857559836
290
+ ],
291
+ [
292
+ 61.52302247518973,
293
+ 72.9452714535902,
294
+ 84.62728035610041
295
+ ],
296
+ [
297
+ 62.357304436660826,
298
+ 74.39871570344425,
299
+ 86.16918417980152
300
+ ],
301
+ [
302
+ 62.89430093403386,
303
+ 75.2756129597198,
304
+ 87.13600043782837
305
+ ],
306
+ [
307
+ 63.096559398715705,
308
+ 75.58358873321657,
309
+ 87.46008464681844
310
+ ]
311
+ ],
312
+ "l0_aligned_by_layer": [
313
+ [
314
+ 108.76663747810858,
315
+ 116.1836325160537,
316
+ 129.81337565674255
317
+ ],
318
+ [
319
+ 59.149591360186804,
320
+ 69.2851539696439,
321
+ 81.08269483362523
322
+ ],
323
+ [
324
+ 58.44850043782837,
325
+ 68.96836690017513,
326
+ 80.13771526561588
327
+ ],
328
+ [
329
+ 59.264685493286635,
330
+ 70.46750948628137,
331
+ 81.48279699357852
332
+ ],
333
+ [
334
+ 60.21510143023934,
335
+ 72.08517586106247,
336
+ 83.3359420607122
337
+ ],
338
+ [
339
+ 60.98376386456509,
340
+ 73.48206727962639,
341
+ 84.82963003502627
342
+ ],
343
+ [
344
+ 61.47834573847052,
345
+ 74.30782983070637,
346
+ 85.76492265032107
347
+ ],
348
+ [
349
+ 61.64835084646818,
350
+ 74.61715557501459,
351
+ 86.1044038237011
352
+ ]
353
+ ],
354
+ "val_l0_base": [
355
+ 74.9518095485827,
356
+ 72.04285087385726,
357
+ 71.85811411273417,
358
+ 73.11738550475755,
359
+ 74.9007987376907,
360
+ 75.9018712867617,
361
+ 76.71053545886933,
362
+ 76.88609194131422
363
+ ],
364
+ "val_l0_aligned": [
365
+ 73.2166620624003,
366
+ 70.44959171155361,
367
+ 70.45259094238281,
368
+ 72.05052217014173,
369
+ 73.63809156168193,
370
+ 74.55457613480652,
371
+ 75.40545250857687,
372
+ 75.5361974426589
373
+ ],
374
+ "val_l0_base_by_layer": [
375
+ [
376
+ 63.72376746532181,
377
+ 74.18455497382199,
378
+ 86.94709861715427
379
+ ],
380
+ [
381
+ 61.39261563286107,
382
+ 71.40614095158602,
383
+ 83.32978840773018
384
+ ],
385
+ [
386
+ 61.2749781983061,
387
+ 70.88192629190016,
388
+ 83.41743021860172
389
+ ],
390
+ [
391
+ 61.58027924542652,
392
+ 73.10269416689249,
393
+ 84.66917539267016
394
+ ],
395
+ [
396
+ 62.88618019862949,
397
+ 75.00589005235602,
398
+ 86.81031851244222
399
+ ],
400
+ [
401
+ 63.25599914071447,
402
+ 76.11076570680628,
403
+ 88.33884162303664
404
+ ],
405
+ [
406
+ 64.0468477378965,
407
+ 76.84511346467502,
408
+ 89.23963788416998
409
+ ],
410
+ [
411
+ 64.24105585307976,
412
+ 77.04439356938707,
413
+ 89.37281851244222
414
+ ]
415
+ ],
416
+ "val_l0_aligned_by_layer": [
417
+ [
418
+ 62.179482984293195,
419
+ 72.61747382198953,
420
+ 84.85302139202338
421
+ ],
422
+ [
423
+ 59.61180192637818,
424
+ 70.26423429319372,
425
+ 81.47273126572215
426
+ ],
427
+ [
428
+ 59.49803664921466,
429
+ 70.10842060668307,
430
+ 81.75130890052355
431
+ ],
432
+ [
433
+ 60.45980585307975,
434
+ 72.25725351703105,
435
+ 83.434500449615
436
+ ],
437
+ [
438
+ 61.48102094240838,
439
+ 73.94295377881114,
440
+ 85.49029233443176
441
+ ],
442
+ [
443
+ 61.75714442867259,
444
+ 74.9690227009239,
445
+ 86.93755456415147
446
+ ],
447
+ [
448
+ 62.58933246073298,
449
+ 75.80775523560209,
450
+ 87.81926267933471
451
+ ],
452
+ [
453
+ 62.69873474280872,
454
+ 75.9856566583923,
455
+ 87.92419285799197
456
+ ]
457
+ ],
458
+ "self_recon": [
459
+ 0.03777978854700793,
460
+ 0.027052706861036672,
461
+ 0.025448514492198242,
462
+ 0.02454077672176346,
463
+ 0.02385480677675275,
464
+ 0.02330240188954834,
465
+ 0.022913188621906323,
466
+ 0.02270352855320396
467
+ ],
468
+ "cross_recon": [
469
+ 0.03782143883333176,
470
+ 0.027059119352413005,
471
+ 0.025455742043601234,
472
+ 0.02455789306061315,
473
+ 0.02389682638331143,
474
+ 0.023368101301714475,
475
+ 0.022991156307746636,
476
+ 0.022785278416678562
477
+ ],
478
+ "sparsity": [
479
+ 0.023822688434807986,
480
+ 0.005265172788959216,
481
+ 0.0054389986500983445,
482
+ 0.005584087701795989,
483
+ 0.0057119209736898775,
484
+ 0.005816767561171689,
485
+ 0.005884397146713382,
486
+ 0.005909877130625823
487
+ ],
488
+ "val_self_recon": [
489
+ 0.028549473235354374,
490
+ 0.026422248499867802,
491
+ 0.025580557531321237,
492
+ 0.025051593956054816,
493
+ 0.024642081082803417,
494
+ 0.02439956735644041,
495
+ 0.024270332832845094,
496
+ 0.024245827668468365
497
+ ],
498
+ "val_cross_recon": [
499
+ 0.028555416477678335,
500
+ 0.02642720296042752,
501
+ 0.02558600322309277,
502
+ 0.02506005116470197,
503
+ 0.024656025887861924,
504
+ 0.0244177212324286,
505
+ 0.024290522209402778,
506
+ 0.02426636498909034
507
+ ],
508
+ "val_sparsity": [
509
+ 0.005231669019049531,
510
+ 0.005410865241788445,
511
+ 0.0054814473564946216,
512
+ 0.005573431137188567,
513
+ 0.00567762855274593,
514
+ 0.005746121440297334,
515
+ 0.0058087650418320565,
516
+ 0.0058244062753404
517
+ ],
518
+ "layers": [
519
+ 12,
520
+ 13,
521
+ 14
522
+ ],
523
+ "topk_mode": "model_balanced_layer_agg"
524
+ }
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/aligned_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: 0b785c0789235e0e09021152111abe2883320037492c285cb1cf161898bf4426
  • Pointer size: 131 Bytes
  • Size of remote file: 194 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/base_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: 7963e169c56dc9770c48ebefcc48cf8e0fb74093580fa70215efb62ed9e1325b
  • Pointer size: 131 Bytes
  • Size of remote file: 193 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/cf_shift_by_layer.png ADDED

Git LFS Details

  • SHA256: 5bb386efba3e01de7132c86769c5ab2467c86e965676e39b6804d118f6f6e947
  • Pointer size: 130 Bytes
  • Size of remote file: 44 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/cf_shift_p95_by_layer.png ADDED

Git LFS Details

  • SHA256: 1b98235767fb49dc13db68afb0d9b4c78927e833808be066213cd3f90714e32d
  • Pointer size: 131 Bytes
  • Size of remote file: 116 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/class_distribution_multilayer.png ADDED

Git LFS Details

  • SHA256: ff6c710cc3fb0f7f7b4dd3b91cbd73c7516f52833d1434f366486918b81e7ccf
  • Pointer size: 130 Bytes
  • Size of remote file: 83.8 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/class_distribution_primary.png ADDED

Git LFS Details

  • SHA256: 6450e25fffc381fdd660003367ca532e23cb407b552824d0810da1e6cbb5e414
  • Pointer size: 131 Bytes
  • Size of remote file: 146 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/cross_layer_cosine_drift_by_stream.png ADDED

Git LFS Details

  • SHA256: d9a5495cabc23385c43815cd409e968b6155c052712e4add89302f66e868c10e
  • Pointer size: 130 Bytes
  • Size of remote file: 54.7 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/decoder_norm_ratio_by_layer.png ADDED

Git LFS Details

  • SHA256: 95a6ee52212b466029370571e30669187732f4a225cf249ee6b5be1eb76614a0
  • Pointer size: 130 Bytes
  • Size of remote file: 50.2 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/feature_layer_trajectories.png ADDED

Git LFS Details

  • SHA256: b50250e7ab757910a1afec29c7b7b91f669a54518082bb8d0fefd85b0f0c78ac
  • Pointer size: 131 Bytes
  • Size of remote file: 609 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/feature_sharing_ratio_by_layer.png ADDED

Git LFS Details

  • SHA256: 79372f07321e52568481e022c394f15359ead5e999bd541279f0931175d2ac69
  • Pointer size: 130 Bytes
  • Size of remote file: 62.3 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/fve_by_layer.png ADDED

Git LFS Details

  • SHA256: b55d5c6286ab990ae0871f45d91ea5ab804868ca6b5f4fd4f7369462e97fa405
  • Pointer size: 131 Bytes
  • Size of remote file: 104 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/l0_by_layer.png ADDED

Git LFS Details

  • SHA256: 705f648cc6291e9ef79862fdf65a3fb9e6bc7640e8de25d608d71dff7476b781
  • Pointer size: 130 Bytes
  • Size of remote file: 91.6 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/layer_concentration_entropy.png ADDED

Git LFS Details

  • SHA256: 312d9cb7fa53e91947d5e5858294916964b0595d05ac8c0e56b68203c824cd5a
  • Pointer size: 131 Bytes
  • Size of remote file: 107 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/loss_curves.png ADDED

Git LFS Details

  • SHA256: c4e6fa4dfe3a8537641f5d3139ce0130686b2554b04827d8bffbcecc19a5739d
  • Pointer size: 131 Bytes
  • Size of remote file: 327 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/max_norm_layer_migration.png ADDED

Git LFS Details

  • SHA256: 87135a29f45f35bbe7cdc9119b42d5471a261be7b413c570f74bb2a0ff50f1a6
  • Pointer size: 130 Bytes
  • Size of remote file: 80.1 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/rho_histogram_by_layer.png ADDED

Git LFS Details

  • SHA256: fcebb92a72b5be882bcfd1f347ec756a05a1019895002a2e43b7fac4a9a5825a
  • Pointer size: 130 Bytes
  • Size of remote file: 55.3 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/rho_theta_scatter_by_layer.png ADDED

Git LFS Details

  • SHA256: ee8823cc50c293947c198259274ddad737015da8bff88e8d6e095f617e539d5d
  • Pointer size: 132 Bytes
  • Size of remote file: 1.17 MB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/superposition_by_layer.png ADDED

Git LFS Details

  • SHA256: fc8f4a0305ed1c9069347275b1244a1f8a55ed5a755eb3d3a261ac501c52b2a0
  • Pointer size: 131 Bytes
  • Size of remote file: 107 kB
v1-undetermined/assembled_activations/llama32-3b-dpo/plots/theta_by_layer.png ADDED

Git LFS Details

  • SHA256: 1553d348e4ea95b02e7551fd2cd2d9f6fb83b59cf8d0b0d5b0adfe5d2c8610d3
  • Pointer size: 130 Bytes
  • Size of remote file: 51.6 kB
v1-undetermined/assembled_activations/llama32-3b-grpo/checkpoints/final.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab5c226343b2042cb69e8cba11d4f0ff6753f3d104d3d8a1a036e8e217e58153
3
+ size 10873439922
v1-undetermined/assembled_activations/llama32-3b-grpo/features/counterfactual_scores.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-grpo/features/counterfactual_scores_by_layer.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-grpo/features/cross_layer_cosine_drift.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da7045e64bece542f6164c96fe5d4646c516a3f92c58101857158273d5197575
3
+ size 27244726
v1-undetermined/assembled_activations/llama32-3b-grpo/features/decoder_layer_profiles.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7e8c1f429a3bfda239e53e7c0638b75f2511ed00bf7556ed10bed296bb20867
3
+ size 13037100
v1-undetermined/assembled_activations/llama32-3b-grpo/features/feature_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-grpo/features/merged_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-grpo/features/model_layer_stream_patterns.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-grpo/features/superposition_analysis.json ADDED
The diff for this file is too large to render. See raw diff
 
v1-undetermined/assembled_activations/llama32-3b-grpo/metrics/aggregate_metrics.json ADDED
@@ -0,0 +1,388 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crosscoder_kind": "multilayer_sparc",
3
+ "layers": [
4
+ 12,
5
+ 13,
6
+ 14
7
+ ],
8
+ "topk_mode": "model_balanced_layer_agg",
9
+ "topk": 400,
10
+ "expansion_factor": 8,
11
+ "dict_size": 24576,
12
+ "forced_shared_fraction": 0.06,
13
+ "class_counts": {
14
+ "aligned_only": 6084,
15
+ "shared_redirected": 5805,
16
+ "shared_intermediate": 4922,
17
+ "base_only": 3467,
18
+ "shared_attenuated": 2812,
19
+ "shared_aligned": 1486
20
+ },
21
+ "multilayer_class_counts": {
22
+ "drifting_or_rotating": 12498,
23
+ "persistent_aligned_only": 4582,
24
+ "persistent_base_only": 2198,
25
+ "persistent_shared": 2170,
26
+ "localized_aligned_only": 1502,
27
+ "localized_base_only": 1269,
28
+ "mixed_or_ambiguous": 357
29
+ },
30
+ "classification_thresholds": {
31
+ "rho_base_only": 0.4,
32
+ "rho_aligned_only": 0.5552206029447704,
33
+ "rho_shared_low": 0.44024917884104786,
34
+ "rho_shared_high": 0.5552206029447704
35
+ },
36
+ "threshold_sensitivity": {
37
+ "original": {
38
+ "aligned_only": 6084,
39
+ "shared_redirected": 5805,
40
+ "shared_intermediate": 4922,
41
+ "base_only": 3467,
42
+ "shared_attenuated": 2812,
43
+ "shared_aligned": 1486
44
+ },
45
+ "perturbed": {
46
+ "delta_-0.05": {
47
+ "base_only": 878,
48
+ "aligned_only": 2725,
49
+ "shared_aligned": 1499,
50
+ "shared_redirected": 9722,
51
+ "shared_intermediate": 7743,
52
+ "shared_attenuated": 2009,
53
+ "other": 0
54
+ },
55
+ "delta_+0.05": {
56
+ "base_only": 7128,
57
+ "aligned_only": 10726,
58
+ "shared_aligned": 1475,
59
+ "shared_redirected": 771,
60
+ "shared_intermediate": 646,
61
+ "shared_attenuated": 3830,
62
+ "other": 0
63
+ }
64
+ },
65
+ "perturbation": 0.05
66
+ },
67
+ "class_counts_by_layer": {
68
+ "12": {
69
+ "shared_redirected": 8317,
70
+ "shared_intermediate": 6544,
71
+ "base_only": 4585,
72
+ "shared_aligned": 2822,
73
+ "aligned_only": 2308
74
+ },
75
+ "13": {
76
+ "shared_redirected": 9393,
77
+ "shared_intermediate": 8081,
78
+ "base_only": 3041,
79
+ "shared_aligned": 2669,
80
+ "aligned_only": 1392
81
+ },
82
+ "14": {
83
+ "shared_redirected": 9247,
84
+ "base_only": 4609,
85
+ "shared_intermediate": 4229,
86
+ "aligned_only": 2292,
87
+ "shared_aligned": 2107,
88
+ "shared_attenuated": 2092
89
+ }
90
+ },
91
+ "feature_sharing_ratio_by_layer": {
92
+ "12": 0.7195231119791666,
93
+ "13": 0.8196207682291666,
94
+ "14": 0.7191975911458334
95
+ },
96
+ "decoder_amplification_by_layer": {
97
+ "12": {
98
+ "median": 0.9999999814920351,
99
+ "p95": 4.761198471404814
100
+ },
101
+ "13": {
102
+ "median": 0.9999999845770899,
103
+ "p95": 2.7799582920911448
104
+ },
105
+ "14": {
106
+ "median": 0.9999999870003597,
107
+ "p95": 1.9990551070509737
108
+ }
109
+ },
110
+ "classification_thresholds_by_layer": {
111
+ "12": {
112
+ "rho_base_only": 0.3232246755639113,
113
+ "rho_aligned_only": 0.7496195037077811,
114
+ "rho_shared_low": 0.3232246755639113,
115
+ "rho_shared_high": 0.7496195037077811
116
+ },
117
+ "13": {
118
+ "rho_base_only": 0.33013993411090486,
119
+ "rho_aligned_only": 0.7235388298969397,
120
+ "rho_shared_low": 0.33013993411090486,
121
+ "rho_shared_high": 0.7235388298969397
122
+ },
123
+ "14": {
124
+ "rho_base_only": 0.4,
125
+ "rho_aligned_only": 0.6325617662750056,
126
+ "rho_shared_low": 0.43268800458718787,
127
+ "rho_shared_high": 0.6325617662750056
128
+ }
129
+ },
130
+ "threshold_sensitivity_by_layer": {
131
+ "12": {
132
+ "original": {
133
+ "shared_redirected": 8317,
134
+ "shared_intermediate": 6544,
135
+ "base_only": 4585,
136
+ "shared_aligned": 2822,
137
+ "aligned_only": 2308
138
+ },
139
+ "perturbed": {
140
+ "delta_-0.05": {
141
+ "base_only": 3098,
142
+ "aligned_only": 1560,
143
+ "shared_aligned": 3403,
144
+ "shared_redirected": 9201,
145
+ "shared_intermediate": 7314,
146
+ "shared_attenuated": 0,
147
+ "other": 0
148
+ },
149
+ "delta_+0.05": {
150
+ "base_only": 6526,
151
+ "aligned_only": 3453,
152
+ "shared_aligned": 2068,
153
+ "shared_redirected": 7069,
154
+ "shared_intermediate": 5460,
155
+ "shared_attenuated": 0,
156
+ "other": 0
157
+ }
158
+ },
159
+ "perturbation": 0.05
160
+ },
161
+ "13": {
162
+ "original": {
163
+ "shared_redirected": 9393,
164
+ "shared_intermediate": 8081,
165
+ "base_only": 3041,
166
+ "shared_aligned": 2669,
167
+ "aligned_only": 1392
168
+ },
169
+ "perturbed": {
170
+ "delta_-0.05": {
171
+ "base_only": 1800,
172
+ "aligned_only": 845,
173
+ "shared_aligned": 3022,
174
+ "shared_redirected": 10094,
175
+ "shared_intermediate": 8815,
176
+ "shared_attenuated": 0,
177
+ "other": 0
178
+ },
179
+ "delta_+0.05": {
180
+ "base_only": 5121,
181
+ "aligned_only": 2474,
182
+ "shared_aligned": 2031,
183
+ "shared_redirected": 8218,
184
+ "shared_intermediate": 6732,
185
+ "shared_attenuated": 0,
186
+ "other": 0
187
+ }
188
+ },
189
+ "perturbation": 0.05
190
+ },
191
+ "14": {
192
+ "original": {
193
+ "shared_redirected": 9247,
194
+ "base_only": 4609,
195
+ "shared_intermediate": 4229,
196
+ "aligned_only": 2292,
197
+ "shared_aligned": 2107,
198
+ "shared_attenuated": 2092
199
+ },
200
+ "perturbed": {
201
+ "delta_-0.05": {
202
+ "base_only": 2008,
203
+ "aligned_only": 921,
204
+ "shared_aligned": 2965,
205
+ "shared_redirected": 11424,
206
+ "shared_intermediate": 5725,
207
+ "shared_attenuated": 1533,
208
+ "other": 0
209
+ },
210
+ "delta_+0.05": {
211
+ "base_only": 7950,
212
+ "aligned_only": 5052,
213
+ "shared_aligned": 1379,
214
+ "shared_redirected": 5406,
215
+ "shared_intermediate": 2165,
216
+ "shared_attenuated": 2624,
217
+ "other": 0
218
+ }
219
+ },
220
+ "perturbation": 0.05
221
+ }
222
+ },
223
+ "counterfactual_shift_by_layer": {
224
+ "12": {
225
+ "aligned_only": {
226
+ "mean_shift": 2.2766621525107953e-06,
227
+ "median_shift": 0.0,
228
+ "p95_abs_shift": 0.0,
229
+ "count": 6084
230
+ },
231
+ "base_only": {
232
+ "mean_shift": -2.568851394574549e-06,
233
+ "median_shift": 0.0,
234
+ "p95_abs_shift": 0.0,
235
+ "count": 3467
236
+ },
237
+ "shared_aligned": {
238
+ "mean_shift": 0.00018095357842273262,
239
+ "median_shift": 0.0,
240
+ "p95_abs_shift": 0.09195232205092907,
241
+ "count": 1486
242
+ },
243
+ "shared_attenuated": {
244
+ "mean_shift": -1.0541640883041274e-07,
245
+ "median_shift": 0.0,
246
+ "p95_abs_shift": 0.0,
247
+ "count": 2812
248
+ },
249
+ "shared_intermediate": {
250
+ "mean_shift": 2.315150216625542e-08,
251
+ "median_shift": 0.0,
252
+ "p95_abs_shift": 0.0,
253
+ "count": 4922
254
+ },
255
+ "shared_redirected": {
256
+ "mean_shift": -1.3746285198566602e-07,
257
+ "median_shift": 0.0,
258
+ "p95_abs_shift": 0.0,
259
+ "count": 5805
260
+ }
261
+ },
262
+ "13": {
263
+ "aligned_only": {
264
+ "mean_shift": 7.409373586140498e-06,
265
+ "median_shift": 0.0,
266
+ "p95_abs_shift": 0.0,
267
+ "count": 6084
268
+ },
269
+ "base_only": {
270
+ "mean_shift": -1.1053492490492892e-05,
271
+ "median_shift": 0.0,
272
+ "p95_abs_shift": 0.0,
273
+ "count": 3467
274
+ },
275
+ "shared_aligned": {
276
+ "mean_shift": 0.00019484710632688615,
277
+ "median_shift": 0.0,
278
+ "p95_abs_shift": 0.09973335079848766,
279
+ "count": 1486
280
+ },
281
+ "shared_attenuated": {
282
+ "mean_shift": -1.985480392448661e-06,
283
+ "median_shift": 0.0,
284
+ "p95_abs_shift": 0.0,
285
+ "count": 2812
286
+ },
287
+ "shared_intermediate": {
288
+ "mean_shift": 3.9290498327251884e-08,
289
+ "median_shift": 0.0,
290
+ "p95_abs_shift": 0.0,
291
+ "count": 4922
292
+ },
293
+ "shared_redirected": {
294
+ "mean_shift": -1.7551452070176433e-07,
295
+ "median_shift": 0.0,
296
+ "p95_abs_shift": 0.0,
297
+ "count": 5805
298
+ }
299
+ },
300
+ "14": {
301
+ "aligned_only": {
302
+ "mean_shift": 3.364509558595365e-05,
303
+ "median_shift": 0.0,
304
+ "p95_abs_shift": 0.0,
305
+ "count": 6084
306
+ },
307
+ "base_only": {
308
+ "mean_shift": -3.5504677135829413e-05,
309
+ "median_shift": 0.0,
310
+ "p95_abs_shift": 0.0,
311
+ "count": 3467
312
+ },
313
+ "shared_aligned": {
314
+ "mean_shift": 0.0003855111859048331,
315
+ "median_shift": 0.0,
316
+ "p95_abs_shift": 0.11618590541183949,
317
+ "count": 1486
318
+ },
319
+ "shared_attenuated": {
320
+ "mean_shift": -9.63817584883231e-06,
321
+ "median_shift": 0.0,
322
+ "p95_abs_shift": 0.0,
323
+ "count": 2812
324
+ },
325
+ "shared_intermediate": {
326
+ "mean_shift": 2.184999210256547e-07,
327
+ "median_shift": 0.0,
328
+ "p95_abs_shift": 0.0,
329
+ "count": 4922
330
+ },
331
+ "shared_redirected": {
332
+ "mean_shift": 4.39443286374644e-07,
333
+ "median_shift": 0.0,
334
+ "p95_abs_shift": 0.0,
335
+ "count": 5805
336
+ }
337
+ }
338
+ },
339
+ "total_features": 24576,
340
+ "fve_base": 0.6947082683678073,
341
+ "fve_aligned": 0.6870226214069346,
342
+ "fve_base_by_layer": [
343
+ 0.6774334142969541,
344
+ 0.6886444531810221,
345
+ 0.7180468839500587
346
+ ],
347
+ "fve_aligned_by_layer": [
348
+ 0.6701876355715447,
349
+ 0.6808480033699754,
350
+ 0.7100321675470362
351
+ ],
352
+ "val_fve_base_by_layer": [
353
+ 0.6774334142969541,
354
+ 0.6886444531810221,
355
+ 0.7180468839500587
356
+ ],
357
+ "val_fve_aligned_by_layer": [
358
+ 0.6701876355715447,
359
+ 0.6808480033699754,
360
+ 0.7100321675470362
361
+ ],
362
+ "dead_neuron_fraction": 0.9609958831845009,
363
+ "l0_sparsity_base": 77.30068727438727,
364
+ "l0_sparsity_aligned": 76.21462362437518,
365
+ "l0_base_by_layer": [
366
+ 65.64209354932866,
367
+ 76.70472124927029,
368
+ 89.5552393461763
369
+ ],
370
+ "l0_aligned_by_layer": [
371
+ 64.81115002918855,
372
+ 75.33169147694105,
373
+ 88.50102159953299
374
+ ],
375
+ "val_l0_base": 78.84989721487953,
376
+ "val_l0_aligned": 77.61215010238567,
377
+ "val_l0_base_by_layer": [
378
+ 66.79128493064361,
379
+ 78.28724914071448,
380
+ 91.47114966427469
381
+ ],
382
+ "val_l0_aligned_by_layer": [
383
+ 65.80791884816755,
384
+ 76.6645397166307,
385
+ 90.36398343390819
386
+ ],
387
+ "superposition_fraction": 0.0
388
+ }
v1-undetermined/assembled_activations/llama32-3b-grpo/metrics/training_metrics.json ADDED
@@ -0,0 +1,524 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epochs": [
3
+ 1,
4
+ 2,
5
+ 3,
6
+ 4,
7
+ 5,
8
+ 6,
9
+ 7,
10
+ 8
11
+ ],
12
+ "train_loss": [
13
+ 0.07674359122668521,
14
+ 0.04322117879786689,
15
+ 0.041181389142373674,
16
+ 0.040074929835299825,
17
+ 0.0392595374999134,
18
+ 0.03860885386235068,
19
+ 0.038139252809800045,
20
+ 0.03787409050974036
21
+ ],
22
+ "val_loss": [
23
+ 0.045276696184230726,
24
+ 0.042513850998347966,
25
+ 0.04139913450829021,
26
+ 0.04075757585738966,
27
+ 0.04028706457368366,
28
+ 0.0400088208079026,
29
+ 0.039891897734859225,
30
+ 0.03987071945951247
31
+ ],
32
+ "train_fve_base": [
33
+ 0.5218191150209165,
34
+ 0.6581393129562538,
35
+ 0.6788746238589913,
36
+ 0.6907193261979967,
37
+ 0.6996899498740763,
38
+ 0.7067874284146894,
39
+ 0.7117742059248898,
40
+ 0.7144628610794682
41
+ ],
42
+ "train_fve_aligned": [
43
+ 0.5172356182036669,
44
+ 0.6521651021444651,
45
+ 0.6710452417915354,
46
+ 0.6816653398067298,
47
+ 0.6897432762005145,
48
+ 0.6963027055612866,
49
+ 0.7009537841100408,
50
+ 0.7034451776753374
51
+ ],
52
+ "val_fve_base": [
53
+ 0.6393531760620197,
54
+ 0.6663743301211852,
55
+ 0.6773919553656853,
56
+ 0.6840992772766433,
57
+ 0.6891960114708746,
58
+ 0.6927146209472137,
59
+ 0.6943349950600669,
60
+ 0.6947082683678073
61
+ ],
62
+ "val_fve_aligned": [
63
+ 0.634585750664716,
64
+ 0.660018102348787,
65
+ 0.6706369740800707,
66
+ 0.6769087074314737,
67
+ 0.68184863584828,
68
+ 0.6851667025326434,
69
+ 0.6866701738372523,
70
+ 0.6870226214069346
71
+ ],
72
+ "val_fve_base_by_layer": [
73
+ [
74
+ 0.6272152599863982,
75
+ 0.6332998275756836,
76
+ 0.6575443794590017
77
+ ],
78
+ [
79
+ 0.6521256428738539,
80
+ 0.6600078350586417,
81
+ 0.686989458755673
82
+ ],
83
+ [
84
+ 0.6621782736004335,
85
+ 0.6711501657026601,
86
+ 0.6988473628203906
87
+ ],
88
+ [
89
+ 0.6676703444950244,
90
+ 0.6779192161809712,
91
+ 0.7067082190388785
92
+ ],
93
+ [
94
+ 0.6722474247997344,
95
+ 0.6831871764198023,
96
+ 0.7121533692195153
97
+ ],
98
+ [
99
+ 0.6756168560207826,
100
+ 0.6865271073985474,
101
+ 0.7159998494917186
102
+ ],
103
+ [
104
+ 0.6770371470151771,
105
+ 0.6882066729805232,
106
+ 0.7177611193107685
107
+ ],
108
+ [
109
+ 0.6774334142969541,
110
+ 0.6886444531810221,
111
+ 0.7180468839500587
112
+ ]
113
+ ],
114
+ "val_fve_aligned_by_layer": [
115
+ [
116
+ 0.6233792255062083,
117
+ 0.6277059453944261,
118
+ 0.6526720202406039
119
+ ],
120
+ [
121
+ 0.6468782418685434,
122
+ 0.653206247933872,
123
+ 0.679969761072029
124
+ ],
125
+ [
126
+ 0.6562783268733798,
127
+ 0.6638215808344137,
128
+ 0.6918109536795092
129
+ ],
130
+ [
131
+ 0.6612628511109278,
132
+ 0.6705732148979346,
133
+ 0.6988899854465305
134
+ ],
135
+ [
136
+ 0.6654992702743769,
137
+ 0.6759732140920549,
138
+ 0.7040733623254986
139
+ ],
140
+ [
141
+ 0.6685321296696888,
142
+ 0.6789075569956714,
143
+ 0.7080603703778452
144
+ ],
145
+ [
146
+ 0.6698048367550236,
147
+ 0.6804219499932533,
148
+ 0.7097836757829676
149
+ ],
150
+ [
151
+ 0.6701876355715447,
152
+ 0.6808480033699754,
153
+ 0.7100321675470362
154
+ ]
155
+ ],
156
+ "train_fve_base_by_layer": [
157
+ [
158
+ 0.513117791544396,
159
+ 0.5155797629278442,
160
+ 0.5367597415999787
161
+ ],
162
+ [
163
+ 0.6441039162216587,
164
+ 0.6522654140823707,
165
+ 0.6780485487513144
166
+ ],
167
+ [
168
+ 0.6628799028933848,
169
+ 0.6730503911381914,
170
+ 0.7006935184626849
171
+ ],
172
+ [
173
+ 0.6733515275478085,
174
+ 0.685189020821593,
175
+ 0.7136173695760795
176
+ ],
177
+ [
178
+ 0.6809947529663068,
179
+ 0.6943080470148612,
180
+ 0.7237669889577564
181
+ ],
182
+ [
183
+ 0.6869359059912797,
184
+ 0.7013317935312274,
185
+ 0.7320945297704409
186
+ ],
187
+ [
188
+ 0.691114535056146,
189
+ 0.7062968335789126,
190
+ 0.7379111921794221
191
+ ],
192
+ [
193
+ 0.6934602800267382,
194
+ 0.7089756050794638,
195
+ 0.74095263758808
196
+ ]
197
+ ],
198
+ "train_fve_aligned_by_layer": [
199
+ [
200
+ 0.5089469564914425,
201
+ 0.510429343701244,
202
+ 0.5323305079203493
203
+ ],
204
+ [
205
+ 0.639565573195164,
206
+ 0.6456975710719076,
207
+ 0.6712321020745424
208
+ ],
209
+ [
210
+ 0.6564315080433928,
211
+ 0.664647905400672,
212
+ 0.6920562531957832
213
+ ],
214
+ [
215
+ 0.6654149732322493,
216
+ 0.6757753993500344,
217
+ 0.7038055851455328
218
+ ],
219
+ [
220
+ 0.6721183056032456,
221
+ 0.6841006580668447,
222
+ 0.7130108071013772
223
+ ],
224
+ [
225
+ 0.6774491685841143,
226
+ 0.6906084561306207,
227
+ 0.7208504371662552
228
+ ],
229
+ [
230
+ 0.6813007789749328,
231
+ 0.6952133505092472,
232
+ 0.726347162510593
233
+ ],
234
+ [
235
+ 0.6834543206437039,
236
+ 0.6976807647049253,
237
+ 0.729200392039422
238
+ ]
239
+ ],
240
+ "dead_neurons": [
241
+ 0.8841513978133616,
242
+ 0.9568707898076486,
243
+ 0.9615639534110711,
244
+ 0.9622459215026673,
245
+ 0.961871207317134,
246
+ 0.9613948494665677,
247
+ 0.9610488299939345,
248
+ 0.9609958831845009
249
+ ],
250
+ "l0_base": [
251
+ 121.136466098124,
252
+ 72.97896241891196,
253
+ 72.2001569606516,
254
+ 73.32550964979004,
255
+ 74.82235363249325,
256
+ 76.16007140917313,
257
+ 77.00653347409543,
258
+ 77.30068727438727
259
+ ],
260
+ "l0_aligned": [
261
+ 118.93421989481003,
262
+ 71.53167821925354,
263
+ 71.0108692305031,
264
+ 72.31692343431818,
265
+ 73.82103408127588,
266
+ 75.11973643929235,
267
+ 75.93497887481672,
268
+ 76.21462362437518
269
+ ],
270
+ "l0_base_by_layer": [
271
+ [
272
+ 111.43872227086982,
273
+ 118.30589608873322,
274
+ 133.66476941039113
275
+ ],
276
+ [
277
+ 62.302958990075894,
278
+ 71.08327860478693,
279
+ 85.55064214827787
280
+ ],
281
+ [
282
+ 61.694559982486865,
283
+ 70.65915061295972,
284
+ 84.24675277291301
285
+ ],
286
+ [
287
+ 62.596194541739635,
288
+ 72.23266929363689,
289
+ 85.14765761821366
290
+ ],
291
+ [
292
+ 63.748741243432576,
293
+ 74.00266345592527,
294
+ 86.71564871570344
295
+ ],
296
+ [
297
+ 64.77014010507881,
298
+ 75.4961325160537,
299
+ 88.21393388791594
300
+ ],
301
+ [
302
+ 65.40331290134267,
303
+ 76.38590922358435,
304
+ 89.23037069468768
305
+ ],
306
+ [
307
+ 65.64209354932866,
308
+ 76.70472124927029,
309
+ 89.5552393461763
310
+ ]
311
+ ],
312
+ "l0_aligned_by_layer": [
313
+ [
314
+ 109.56592965557502,
315
+ 115.97254451255108,
316
+ 131.26417469352015
317
+ ],
318
+ [
319
+ 61.350700525394046,
320
+ 69.53880253940456,
321
+ 83.70552393461763
322
+ ],
323
+ [
324
+ 60.85610040863981,
325
+ 69.35692133683597,
326
+ 82.81957822533568
327
+ ],
328
+ [
329
+ 61.81104057209574,
330
+ 71.03205268534734,
331
+ 84.10766929363689
332
+ ],
333
+ [
334
+ 62.96254743140689,
335
+ 72.73571584938703,
336
+ 85.76483143607706
337
+ ],
338
+ [
339
+ 63.978528166958554,
340
+ 74.16905647985989,
341
+ 87.21161704611792
342
+ ],
343
+ [
344
+ 64.58822241681261,
345
+ 75.03416885580852,
346
+ 88.18253794512552
347
+ ],
348
+ [
349
+ 64.81115002918855,
350
+ 75.33169147694105,
351
+ 88.50102159953299
352
+ ]
353
+ ],
354
+ "val_l0_base": [
355
+ 76.48782280727207,
356
+ 73.48593175598464,
357
+ 73.73878610820671,
358
+ 75.04906807145524,
359
+ 76.41334242096747,
360
+ 77.9022164469614,
361
+ 78.6846667384602,
362
+ 78.84989721487953
363
+ ],
364
+ "val_l0_aligned": [
365
+ 75.0534313261821,
366
+ 72.0293619844926,
367
+ 72.63701885283305,
368
+ 73.83349936919687,
369
+ 75.38107467571479,
370
+ 76.78896056170238,
371
+ 77.48427745559452,
372
+ 77.61215010238567
373
+ ],
374
+ "val_l0_base_by_layer": [
375
+ [
376
+ 65.50414485831536,
377
+ 74.47414921465969,
378
+ 89.48516582069597
379
+ ],
380
+ [
381
+ 63.45653360177085,
382
+ 71.41803010471205,
383
+ 85.58322427160452
384
+ ],
385
+ [
386
+ 63.403250449615,
387
+ 72.4886562087773,
388
+ 85.32444371727749
389
+ ],
390
+ [
391
+ 63.711605611271885,
392
+ 74.16361256544502,
393
+ 87.2719786479211
394
+ ],
395
+ [
396
+ 64.96951353862023,
397
+ 75.69464442867259,
398
+ 88.5758617061595
399
+ ],
400
+ [
401
+ 66.02416013547888,
402
+ 77.07302576833995,
403
+ 90.60945680628272
404
+ ],
405
+ [
406
+ 66.63781634425618,
407
+ 78.01167105629806,
408
+ 91.40450482593157
409
+ ],
410
+ [
411
+ 66.79128493064361,
412
+ 78.28724914071448,
413
+ 91.47114966427469
414
+ ]
415
+ ],
416
+ "val_l0_aligned_by_layer": [
417
+ [
418
+ 64.29243019862949,
419
+ 73.13732550655985,
420
+ 87.73053010471205
421
+ ],
422
+ [
423
+ 61.9127945026178,
424
+ 70.0345222513089,
425
+ 84.14076137043419
426
+ ],
427
+ [
428
+ 62.49950916230367,
429
+ 70.98451137043419,
430
+ 84.42702879581152
431
+ ],
432
+ [
433
+ 62.823243905112385,
434
+ 72.5755890052356,
435
+ 86.10165796729282
436
+ ],
437
+ [
438
+ 64.12390925622111,
439
+ 74.49552793652599,
440
+ 87.52377838614099
441
+ ],
442
+ [
443
+ 65.08524214659685,
444
+ 75.61262000917765,
445
+ 89.66901178010471
446
+ ],
447
+ [
448
+ 65.6529777486911,
449
+ 76.40794068980591,
450
+ 90.3919066583923
451
+ ],
452
+ [
453
+ 65.80791884816755,
454
+ 76.6645397166307,
455
+ 90.36398343390819
456
+ ]
457
+ ],
458
+ "self_recon": [
459
+ 0.03779663238511299,
460
+ 0.027047466899888666,
461
+ 0.02546192713786445,
462
+ 0.02456349793710663,
463
+ 0.023879754270594925,
464
+ 0.023328691720962524,
465
+ 0.02293838971385257,
466
+ 0.022728793613364735
467
+ ],
468
+ "cross_recon": [
469
+ 0.03785394466065101,
470
+ 0.027076600934990442,
471
+ 0.025491758700410178,
472
+ 0.0246022383432881,
473
+ 0.023943013571946507,
474
+ 0.02341594385669305,
475
+ 0.02303808376897655,
476
+ 0.02283220853215063
477
+ ],
478
+ "sparsity": [
479
+ 0.02380538061810293,
480
+ 0.005343071211149,
481
+ 0.005522758250330807,
482
+ 0.005670536276316232,
483
+ 0.005802577443560069,
484
+ 0.005913784346878703,
485
+ 0.005985629326738395,
486
+ 0.006012413198138129
487
+ ],
488
+ "val_self_recon": [
489
+ 0.028536553528764486,
490
+ 0.02644324319251857,
491
+ 0.02557803145292854,
492
+ 0.025054661449805605,
493
+ 0.0246538199016249,
494
+ 0.02437961772474319,
495
+ 0.024254350498786773,
496
+ 0.024226347998720813
497
+ ],
498
+ "val_cross_recon": [
499
+ 0.028564947554890398,
500
+ 0.02647020462997921,
501
+ 0.0256076222236868,
502
+ 0.02508811676065335,
503
+ 0.024693033419042356,
504
+ 0.024424381593135015,
505
+ 0.024301334952262683,
506
+ 0.02427353766498141
507
+ ],
508
+ "val_sparsity": [
509
+ 0.005314163556468737,
510
+ 0.0054825257890277505,
511
+ 0.005578053820176118,
512
+ 0.005667667375165599,
513
+ 0.005756030966045033,
514
+ 0.005859450251839236,
515
+ 0.00591701309620866,
516
+ 0.005934955881352668
517
+ ],
518
+ "layers": [
519
+ 12,
520
+ 13,
521
+ 14
522
+ ],
523
+ "topk_mode": "model_balanced_layer_agg"
524
+ }
v1-undetermined/assembled_activations/llama32-3b-grpo/plots/aligned_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: f179fead9df779fc48f1573c2c8935981a29cbfff6f8d8d718ff48aa8df82c89
  • Pointer size: 131 Bytes
  • Size of remote file: 188 kB
v1-undetermined/assembled_activations/llama32-3b-grpo/plots/base_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: bfd400890bbf5fa81872f9292127c79bbe7e37c53425f8f94ddab09c91ffa1af
  • Pointer size: 131 Bytes
  • Size of remote file: 187 kB
v1-undetermined/assembled_activations/llama32-3b-grpo/plots/cf_shift_by_layer.png ADDED

Git LFS Details

  • SHA256: 5bb386efba3e01de7132c86769c5ab2467c86e965676e39b6804d118f6f6e947
  • Pointer size: 130 Bytes
  • Size of remote file: 44 kB
v1-undetermined/assembled_activations/llama32-3b-grpo/plots/cf_shift_p95_by_layer.png ADDED

Git LFS Details

  • SHA256: 2f3b143076d09c6c52194fecb5886c0e0f1e098b84123d39c5c86d8f8cb074f5
  • Pointer size: 131 Bytes
  • Size of remote file: 115 kB
v1-undetermined/assembled_activations/llama32-3b-grpo/plots/class_distribution_multilayer.png ADDED

Git LFS Details

  • SHA256: 7c559df4d6cd1e545ff36ab23fd197f9fa87447acfb963436384f9f623a306b9
  • Pointer size: 130 Bytes
  • Size of remote file: 83.8 kB
v1-undetermined/assembled_activations/llama32-3b-grpo/plots/class_distribution_primary.png ADDED

Git LFS Details

  • SHA256: b2c54e345977052040bd492700c2efbfc3faa119ba84c6cdf0c4909996b99c1a
  • Pointer size: 131 Bytes
  • Size of remote file: 146 kB
v1-undetermined/assembled_activations/llama32-3b-grpo/plots/cross_layer_cosine_drift_by_stream.png ADDED

Git LFS Details

  • SHA256: dba3d52fd2c05d42436711a636ec7fec8017f8330e5f53518818204a15237b9d
  • Pointer size: 130 Bytes
  • Size of remote file: 54.8 kB
v1-undetermined/assembled_activations/llama32-3b-grpo/plots/decoder_norm_ratio_by_layer.png ADDED

Git LFS Details

  • SHA256: 2515fba00de16bea47142354e608726d523e141e83683d72869fbf00e3503fb7
  • Pointer size: 130 Bytes
  • Size of remote file: 48.3 kB