ishangarg183 commited on
Commit
a8ae341
·
verified ·
1 Parent(s): fea5b57

Add assembled activations for llama32-3b-kto

Browse files
Files changed (33) hide show
  1. .gitattributes +2 -0
  2. v1/assembled_activations/llama32-3b-kto/activations/activations.pt +3 -0
  3. v1/assembled_activations/llama32-3b-kto/checkpoints/final.pt +3 -0
  4. v1/assembled_activations/llama32-3b-kto/features/counterfactual_scores.csv +0 -0
  5. v1/assembled_activations/llama32-3b-kto/features/counterfactual_scores_by_layer.csv +0 -0
  6. v1/assembled_activations/llama32-3b-kto/features/cross_layer_cosine_drift.csv +3 -0
  7. v1/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv +3 -0
  8. v1/assembled_activations/llama32-3b-kto/features/feature_activations.pt +3 -0
  9. v1/assembled_activations/llama32-3b-kto/features/feature_classification.csv +0 -0
  10. v1/assembled_activations/llama32-3b-kto/features/merged_classification.csv +0 -0
  11. v1/assembled_activations/llama32-3b-kto/features/model_layer_stream_patterns.csv +0 -0
  12. v1/assembled_activations/llama32-3b-kto/features/superposition_analysis.json +271 -0
  13. v1/assembled_activations/llama32-3b-kto/metrics/aggregate_metrics.json +345 -0
  14. v1/assembled_activations/llama32-3b-kto/metrics/training_metrics.json +524 -0
  15. v1/assembled_activations/llama32-3b-kto/plots/aligned_decoder_norm_heatmap.png +3 -0
  16. v1/assembled_activations/llama32-3b-kto/plots/base_decoder_norm_heatmap.png +3 -0
  17. v1/assembled_activations/llama32-3b-kto/plots/cf_shift_by_layer.png +3 -0
  18. v1/assembled_activations/llama32-3b-kto/plots/cf_shift_p95_by_layer.png +3 -0
  19. v1/assembled_activations/llama32-3b-kto/plots/class_distribution_multilayer.png +3 -0
  20. v1/assembled_activations/llama32-3b-kto/plots/class_distribution_primary.png +3 -0
  21. v1/assembled_activations/llama32-3b-kto/plots/cross_layer_cosine_drift_by_stream.png +3 -0
  22. v1/assembled_activations/llama32-3b-kto/plots/decoder_norm_ratio_by_layer.png +3 -0
  23. v1/assembled_activations/llama32-3b-kto/plots/feature_layer_trajectories.png +3 -0
  24. v1/assembled_activations/llama32-3b-kto/plots/feature_sharing_ratio_by_layer.png +3 -0
  25. v1/assembled_activations/llama32-3b-kto/plots/fve_by_layer.png +3 -0
  26. v1/assembled_activations/llama32-3b-kto/plots/l0_by_layer.png +3 -0
  27. v1/assembled_activations/llama32-3b-kto/plots/layer_concentration_entropy.png +3 -0
  28. v1/assembled_activations/llama32-3b-kto/plots/loss_curves.png +3 -0
  29. v1/assembled_activations/llama32-3b-kto/plots/max_norm_layer_migration.png +3 -0
  30. v1/assembled_activations/llama32-3b-kto/plots/rho_histogram_by_layer.png +3 -0
  31. v1/assembled_activations/llama32-3b-kto/plots/rho_theta_scatter_by_layer.png +3 -0
  32. v1/assembled_activations/llama32-3b-kto/plots/superposition_by_layer.png +3 -0
  33. v1/assembled_activations/llama32-3b-kto/plots/theta_by_layer.png +3 -0
.gitattributes CHANGED
@@ -76,3 +76,5 @@ v1/assembled_activations/llama32-3b-dpo/features/cross_layer_cosine_drift.csv fi
76
  v1/assembled_activations/llama32-3b-dpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
77
  v1/assembled_activations/llama32-3b-grpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
78
  v1/assembled_activations/llama32-3b-grpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
 
 
 
76
  v1/assembled_activations/llama32-3b-dpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
77
  v1/assembled_activations/llama32-3b-grpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
78
  v1/assembled_activations/llama32-3b-grpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
79
+ v1/assembled_activations/llama32-3b-kto/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
80
+ v1/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
v1/assembled_activations/llama32-3b-kto/activations/activations.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:031a8ce6e30f0e94cf173814bde688d7e1879b77987dec414f459a448d62ef6f
3
+ size 4496133925
v1/assembled_activations/llama32-3b-kto/checkpoints/final.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40324b29037cf39bebfb2753779f737db78f0948c0ba9adab6364b25e1cca1ca
3
+ size 10873439922
v1/assembled_activations/llama32-3b-kto/features/counterfactual_scores.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-kto/features/counterfactual_scores_by_layer.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-kto/features/cross_layer_cosine_drift.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67419b2031b8730c4ba88558182d2d40e821df0cadd7fcaca4309cadbec2e92f
3
+ size 26853060
v1/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb7314260fbc192b114f5eede5cf464ace89bafec2a823d2f15081e60fc7a9f1
3
+ size 13167539
v1/assembled_activations/llama32-3b-kto/features/feature_activations.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:366df02792e05c657cc565e2810ae2860e7c5f4994f7d51b32a5dc92ec5e128f
3
+ size 35935153173
v1/assembled_activations/llama32-3b-kto/features/feature_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-kto/features/merged_classification.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-kto/features/model_layer_stream_patterns.csv ADDED
The diff for this file is too large to render. See raw diff
 
v1/assembled_activations/llama32-3b-kto/features/superposition_analysis.json ADDED
@@ -0,0 +1,271 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "analysis_kind": "multilayer_decoder_cosine_candidate_matches",
3
+ "description": "Candidate cross-layer superposition screen: aligned decoder vectors are matched against base decoder vectors across all selected layers/features. This is a geometry-derived screen, not a causal decomposition.",
4
+ "layers": [
5
+ 23,
6
+ 24,
7
+ 25
8
+ ],
9
+ "top_k": 5,
10
+ "cosine_threshold": 0.5,
11
+ "regression_candidates": 256,
12
+ "lasso_alpha": 0.01,
13
+ "n_features_analyzed": 5,
14
+ "features": {
15
+ "2344": {
16
+ "feature_id": 2344,
17
+ "primary_class": "aligned_only",
18
+ "target_layer": 25,
19
+ "r2": -4.07232970456306e-07,
20
+ "n_nonzero": 0,
21
+ "is_superposition": false,
22
+ "constituent_features": [],
23
+ "top_base_matches": [
24
+ {
25
+ "base_feature_id": 24296,
26
+ "base_layer": 24,
27
+ "cosine": 0.45633772015571594,
28
+ "abs_cosine": 0.45633772015571594,
29
+ "is_same_feature": false,
30
+ "is_cross_layer": true
31
+ },
32
+ {
33
+ "base_feature_id": 18687,
34
+ "base_layer": 25,
35
+ "cosine": 0.4250088334083557,
36
+ "abs_cosine": 0.4250088334083557,
37
+ "is_same_feature": false,
38
+ "is_cross_layer": false
39
+ },
40
+ {
41
+ "base_feature_id": 2865,
42
+ "base_layer": 24,
43
+ "cosine": -0.4132736921310425,
44
+ "abs_cosine": 0.4132736921310425,
45
+ "is_same_feature": false,
46
+ "is_cross_layer": true
47
+ },
48
+ {
49
+ "base_feature_id": 7175,
50
+ "base_layer": 23,
51
+ "cosine": -0.39883679151535034,
52
+ "abs_cosine": 0.39883679151535034,
53
+ "is_same_feature": false,
54
+ "is_cross_layer": true
55
+ },
56
+ {
57
+ "base_feature_id": 6201,
58
+ "base_layer": 23,
59
+ "cosine": -0.39130496978759766,
60
+ "abs_cosine": 0.39130496978759766,
61
+ "is_same_feature": false,
62
+ "is_cross_layer": true
63
+ }
64
+ ]
65
+ },
66
+ "11037": {
67
+ "feature_id": 11037,
68
+ "primary_class": "aligned_only",
69
+ "target_layer": 25,
70
+ "r2": -0.0009124147442902686,
71
+ "n_nonzero": 0,
72
+ "is_superposition": false,
73
+ "constituent_features": [],
74
+ "top_base_matches": [
75
+ {
76
+ "base_feature_id": 13612,
77
+ "base_layer": 25,
78
+ "cosine": -0.6919379234313965,
79
+ "abs_cosine": 0.6919379234313965,
80
+ "is_same_feature": false,
81
+ "is_cross_layer": false
82
+ },
83
+ {
84
+ "base_feature_id": 9490,
85
+ "base_layer": 25,
86
+ "cosine": -0.6915594339370728,
87
+ "abs_cosine": 0.6915594339370728,
88
+ "is_same_feature": false,
89
+ "is_cross_layer": false
90
+ },
91
+ {
92
+ "base_feature_id": 13762,
93
+ "base_layer": 25,
94
+ "cosine": 0.6774390935897827,
95
+ "abs_cosine": 0.6774390935897827,
96
+ "is_same_feature": false,
97
+ "is_cross_layer": false
98
+ },
99
+ {
100
+ "base_feature_id": 2287,
101
+ "base_layer": 25,
102
+ "cosine": -0.6305816173553467,
103
+ "abs_cosine": 0.6305816173553467,
104
+ "is_same_feature": false,
105
+ "is_cross_layer": false
106
+ },
107
+ {
108
+ "base_feature_id": 15355,
109
+ "base_layer": 24,
110
+ "cosine": -0.6074807643890381,
111
+ "abs_cosine": 0.6074807643890381,
112
+ "is_same_feature": false,
113
+ "is_cross_layer": true
114
+ }
115
+ ]
116
+ },
117
+ "18301": {
118
+ "feature_id": 18301,
119
+ "primary_class": "aligned_only",
120
+ "target_layer": 23,
121
+ "r2": -4.195340993251584e-05,
122
+ "n_nonzero": 0,
123
+ "is_superposition": false,
124
+ "constituent_features": [],
125
+ "top_base_matches": [
126
+ {
127
+ "base_feature_id": 12929,
128
+ "base_layer": 23,
129
+ "cosine": -0.5700091123580933,
130
+ "abs_cosine": 0.5700091123580933,
131
+ "is_same_feature": false,
132
+ "is_cross_layer": false
133
+ },
134
+ {
135
+ "base_feature_id": 17075,
136
+ "base_layer": 23,
137
+ "cosine": -0.53563392162323,
138
+ "abs_cosine": 0.53563392162323,
139
+ "is_same_feature": false,
140
+ "is_cross_layer": false
141
+ },
142
+ {
143
+ "base_feature_id": 15355,
144
+ "base_layer": 24,
145
+ "cosine": -0.5292654037475586,
146
+ "abs_cosine": 0.5292654037475586,
147
+ "is_same_feature": false,
148
+ "is_cross_layer": true
149
+ },
150
+ {
151
+ "base_feature_id": 6201,
152
+ "base_layer": 23,
153
+ "cosine": -0.5044533014297485,
154
+ "abs_cosine": 0.5044533014297485,
155
+ "is_same_feature": false,
156
+ "is_cross_layer": false
157
+ },
158
+ {
159
+ "base_feature_id": 24296,
160
+ "base_layer": 24,
161
+ "cosine": 0.5009859800338745,
162
+ "abs_cosine": 0.5009859800338745,
163
+ "is_same_feature": false,
164
+ "is_cross_layer": true
165
+ }
166
+ ]
167
+ },
168
+ "19481": {
169
+ "feature_id": 19481,
170
+ "primary_class": "aligned_only",
171
+ "target_layer": 24,
172
+ "r2": -0.00041131219288481624,
173
+ "n_nonzero": 0,
174
+ "is_superposition": false,
175
+ "constituent_features": [],
176
+ "top_base_matches": [
177
+ {
178
+ "base_feature_id": 15355,
179
+ "base_layer": 24,
180
+ "cosine": -0.607896089553833,
181
+ "abs_cosine": 0.607896089553833,
182
+ "is_same_feature": false,
183
+ "is_cross_layer": false
184
+ },
185
+ {
186
+ "base_feature_id": 3311,
187
+ "base_layer": 24,
188
+ "cosine": 0.5702545642852783,
189
+ "abs_cosine": 0.5702545642852783,
190
+ "is_same_feature": false,
191
+ "is_cross_layer": false
192
+ },
193
+ {
194
+ "base_feature_id": 20061,
195
+ "base_layer": 24,
196
+ "cosine": -0.5608834028244019,
197
+ "abs_cosine": 0.5608834028244019,
198
+ "is_same_feature": false,
199
+ "is_cross_layer": false
200
+ },
201
+ {
202
+ "base_feature_id": 16726,
203
+ "base_layer": 24,
204
+ "cosine": -0.5551276206970215,
205
+ "abs_cosine": 0.5551276206970215,
206
+ "is_same_feature": false,
207
+ "is_cross_layer": false
208
+ },
209
+ {
210
+ "base_feature_id": 9490,
211
+ "base_layer": 25,
212
+ "cosine": -0.5516860485076904,
213
+ "abs_cosine": 0.5516860485076904,
214
+ "is_same_feature": false,
215
+ "is_cross_layer": true
216
+ }
217
+ ]
218
+ },
219
+ "24296": {
220
+ "feature_id": 24296,
221
+ "primary_class": "aligned_only",
222
+ "target_layer": 24,
223
+ "r2": -0.00016202182581004365,
224
+ "n_nonzero": 0,
225
+ "is_superposition": false,
226
+ "constituent_features": [],
227
+ "top_base_matches": [
228
+ {
229
+ "base_feature_id": 24296,
230
+ "base_layer": 24,
231
+ "cosine": 0.5111119151115417,
232
+ "abs_cosine": 0.5111119151115417,
233
+ "is_same_feature": true,
234
+ "is_cross_layer": false
235
+ },
236
+ {
237
+ "base_feature_id": 18109,
238
+ "base_layer": 23,
239
+ "cosine": -0.43007758259773254,
240
+ "abs_cosine": 0.43007758259773254,
241
+ "is_same_feature": false,
242
+ "is_cross_layer": true
243
+ },
244
+ {
245
+ "base_feature_id": 7175,
246
+ "base_layer": 23,
247
+ "cosine": -0.42721956968307495,
248
+ "abs_cosine": 0.42721956968307495,
249
+ "is_same_feature": false,
250
+ "is_cross_layer": true
251
+ },
252
+ {
253
+ "base_feature_id": 1879,
254
+ "base_layer": 23,
255
+ "cosine": -0.42496082186698914,
256
+ "abs_cosine": 0.42496082186698914,
257
+ "is_same_feature": false,
258
+ "is_cross_layer": true
259
+ },
260
+ {
261
+ "base_feature_id": 6747,
262
+ "base_layer": 23,
263
+ "cosine": -0.42284414172172546,
264
+ "abs_cosine": 0.42284414172172546,
265
+ "is_same_feature": false,
266
+ "is_cross_layer": true
267
+ }
268
+ ]
269
+ }
270
+ }
271
+ }
v1/assembled_activations/llama32-3b-kto/metrics/aggregate_metrics.json ADDED
@@ -0,0 +1,345 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crosscoder_kind": "multilayer_sparc",
3
+ "layers": [
4
+ 23,
5
+ 24,
6
+ 25
7
+ ],
8
+ "topk_mode": "model_balanced_layer_agg",
9
+ "topk": 400,
10
+ "expansion_factor": 8,
11
+ "dict_size": 24576,
12
+ "forced_shared_fraction": 0.06,
13
+ "class_counts": {
14
+ "shared_redirected": 18354,
15
+ "shared_attenuated": 4083,
16
+ "shared_aligned": 1474,
17
+ "shared_intermediate": 660,
18
+ "aligned_only": 5
19
+ },
20
+ "multilayer_class_counts": {
21
+ "drifting_or_rotating": 22988,
22
+ "persistent_shared": 1388,
23
+ "mixed_or_ambiguous": 195,
24
+ "localized_aligned_only": 5
25
+ },
26
+ "classification_thresholds": {
27
+ "rho_base_only": 0.4,
28
+ "rho_aligned_only": 0.9799827247629397,
29
+ "rho_shared_low": 0.49517253515653925,
30
+ "rho_shared_high": 0.8529490783385397
31
+ },
32
+ "threshold_sensitivity": {
33
+ "original": {
34
+ "shared_redirected": 18354,
35
+ "shared_attenuated": 4083,
36
+ "shared_aligned": 1474,
37
+ "shared_intermediate": 660,
38
+ "aligned_only": 5
39
+ },
40
+ "perturbed": {
41
+ "delta_-0.05": {
42
+ "base_only": 0,
43
+ "aligned_only": 0,
44
+ "shared_aligned": 1474,
45
+ "shared_redirected": 22264,
46
+ "shared_intermediate": 829,
47
+ "shared_attenuated": 4,
48
+ "other": 5
49
+ },
50
+ "delta_+0.05": {
51
+ "base_only": 5,
52
+ "aligned_only": 5,
53
+ "shared_aligned": 0,
54
+ "shared_redirected": 919,
55
+ "shared_intermediate": 43,
56
+ "shared_attenuated": 23604,
57
+ "other": 0
58
+ }
59
+ },
60
+ "perturbation": 0.05
61
+ },
62
+ "class_counts_by_layer": {
63
+ "23": {
64
+ "aligned_only": 24452,
65
+ "base_only": 124
66
+ },
67
+ "24": {
68
+ "aligned_only": 21864,
69
+ "base_only": 2712
70
+ },
71
+ "25": {
72
+ "aligned_only": 24502,
73
+ "base_only": 74
74
+ }
75
+ },
76
+ "feature_sharing_ratio_by_layer": {
77
+ "23": 0.0,
78
+ "24": 0.0,
79
+ "25": 0.0
80
+ },
81
+ "decoder_amplification_by_layer": {
82
+ "23": {
83
+ "median": 1.047627542042361,
84
+ "p95": 1.362319406084207
85
+ },
86
+ "24": {
87
+ "median": 1.0430958416504477,
88
+ "p95": 1.3325615028387257
89
+ },
90
+ "25": {
91
+ "median": 1.0378214066487217,
92
+ "p95": 1.2347764635930154
93
+ }
94
+ },
95
+ "classification_thresholds_by_layer": {
96
+ "23": {
97
+ "rho_base_only": 0.3721634745107997,
98
+ "rho_aligned_only": 0.3721634745107997,
99
+ "rho_shared_low": 0.3721634745107997,
100
+ "rho_shared_high": 0.3721634745107997
101
+ },
102
+ "24": {
103
+ "rho_base_only": 0.46737811492284764,
104
+ "rho_aligned_only": 0.46737811492284764,
105
+ "rho_shared_low": 0.46737811492284764,
106
+ "rho_shared_high": 0.46737811492284764
107
+ },
108
+ "25": {
109
+ "rho_base_only": 0.40263814939318293,
110
+ "rho_aligned_only": 0.40263814939318293,
111
+ "rho_shared_low": 0.40263814939318293,
112
+ "rho_shared_high": 0.40263814939318293
113
+ }
114
+ },
115
+ "threshold_sensitivity_by_layer": {
116
+ "23": {
117
+ "original": {
118
+ "aligned_only": 24452,
119
+ "base_only": 124
120
+ },
121
+ "perturbed": {
122
+ "delta_-0.05": {
123
+ "base_only": 95,
124
+ "aligned_only": 24192,
125
+ "shared_aligned": 43,
126
+ "shared_redirected": 146,
127
+ "shared_intermediate": 100,
128
+ "shared_attenuated": 0,
129
+ "other": 0
130
+ },
131
+ "delta_+0.05": {
132
+ "base_only": 384,
133
+ "aligned_only": 24192,
134
+ "shared_aligned": 0,
135
+ "shared_redirected": 0,
136
+ "shared_intermediate": 0,
137
+ "shared_attenuated": 0,
138
+ "other": 0
139
+ }
140
+ },
141
+ "perturbation": 0.05
142
+ },
143
+ "24": {
144
+ "original": {
145
+ "aligned_only": 21864,
146
+ "base_only": 2712
147
+ },
148
+ "perturbed": {
149
+ "delta_-0.05": {
150
+ "base_only": 280,
151
+ "aligned_only": 10646,
152
+ "shared_aligned": 1374,
153
+ "shared_redirected": 11475,
154
+ "shared_intermediate": 801,
155
+ "shared_attenuated": 0,
156
+ "other": 0
157
+ },
158
+ "delta_+0.05": {
159
+ "base_only": 13930,
160
+ "aligned_only": 10646,
161
+ "shared_aligned": 0,
162
+ "shared_redirected": 0,
163
+ "shared_intermediate": 0,
164
+ "shared_attenuated": 0,
165
+ "other": 0
166
+ }
167
+ },
168
+ "perturbation": 0.05
169
+ },
170
+ "25": {
171
+ "original": {
172
+ "aligned_only": 24502,
173
+ "base_only": 74
174
+ },
175
+ "perturbed": {
176
+ "delta_-0.05": {
177
+ "base_only": 63,
178
+ "aligned_only": 24191,
179
+ "shared_aligned": 22,
180
+ "shared_redirected": 241,
181
+ "shared_intermediate": 59,
182
+ "shared_attenuated": 0,
183
+ "other": 0
184
+ },
185
+ "delta_+0.05": {
186
+ "base_only": 385,
187
+ "aligned_only": 24191,
188
+ "shared_aligned": 0,
189
+ "shared_redirected": 0,
190
+ "shared_intermediate": 0,
191
+ "shared_attenuated": 0,
192
+ "other": 0
193
+ }
194
+ },
195
+ "perturbation": 0.05
196
+ }
197
+ },
198
+ "counterfactual_shift_by_layer": {
199
+ "23": {
200
+ "aligned_only": {
201
+ "mean_shift": 0.05192818641662598,
202
+ "median_shift": 0.0,
203
+ "p95_abs_shift": 2.235345268249511,
204
+ "count": 5
205
+ },
206
+ "shared_aligned": {
207
+ "mean_shift": 0.17152424341133637,
208
+ "median_shift": -7.850458587199682e-06,
209
+ "p95_abs_shift": 4.667830324172954,
210
+ "count": 1474
211
+ },
212
+ "shared_attenuated": {
213
+ "mean_shift": -1.008251538443044e-06,
214
+ "median_shift": 0.0,
215
+ "p95_abs_shift": 0.0,
216
+ "count": 4083
217
+ },
218
+ "shared_intermediate": {
219
+ "mean_shift": 6.39913474182749e-07,
220
+ "median_shift": 0.0,
221
+ "p95_abs_shift": 0.0,
222
+ "count": 660
223
+ },
224
+ "shared_redirected": {
225
+ "mean_shift": 5.061595920287719e-07,
226
+ "median_shift": 0.0,
227
+ "p95_abs_shift": 0.0,
228
+ "count": 18354
229
+ }
230
+ },
231
+ "24": {
232
+ "aligned_only": {
233
+ "mean_shift": 0.4149253680070615,
234
+ "median_shift": 0.0,
235
+ "p95_abs_shift": 9.360336065292357,
236
+ "count": 5
237
+ },
238
+ "shared_aligned": {
239
+ "mean_shift": 0.18706549459255475,
240
+ "median_shift": -1.6757953744672704e-06,
241
+ "p95_abs_shift": 4.258874750137304,
242
+ "count": 1474
243
+ },
244
+ "shared_attenuated": {
245
+ "mean_shift": -7.682575175577572e-07,
246
+ "median_shift": 0.0,
247
+ "p95_abs_shift": 0.0,
248
+ "count": 4083
249
+ },
250
+ "shared_intermediate": {
251
+ "mean_shift": 7.5666555200435615e-06,
252
+ "median_shift": 0.0,
253
+ "p95_abs_shift": 0.0,
254
+ "count": 660
255
+ },
256
+ "shared_redirected": {
257
+ "mean_shift": 6.903093206683238e-07,
258
+ "median_shift": 0.0,
259
+ "p95_abs_shift": 0.0,
260
+ "count": 18354
261
+ }
262
+ },
263
+ "25": {
264
+ "aligned_only": {
265
+ "mean_shift": 0.11894665029558382,
266
+ "median_shift": 8.357907972822431e-06,
267
+ "p95_abs_shift": 2.852829647064209,
268
+ "count": 5
269
+ },
270
+ "shared_aligned": {
271
+ "mean_shift": 0.20790013630214754,
272
+ "median_shift": -8.05746958576492e-06,
273
+ "p95_abs_shift": 4.167022371292107,
274
+ "count": 1474
275
+ },
276
+ "shared_attenuated": {
277
+ "mean_shift": -1.924276880147058e-06,
278
+ "median_shift": 0.0,
279
+ "p95_abs_shift": 0.0,
280
+ "count": 4083
281
+ },
282
+ "shared_intermediate": {
283
+ "mean_shift": 4.6839660514179245e-06,
284
+ "median_shift": 0.0,
285
+ "p95_abs_shift": 0.0,
286
+ "count": 660
287
+ },
288
+ "shared_redirected": {
289
+ "mean_shift": 1.508880238708032e-06,
290
+ "median_shift": 0.0,
291
+ "p95_abs_shift": 0.0,
292
+ "count": 18354
293
+ }
294
+ }
295
+ },
296
+ "total_features": 24576,
297
+ "fve_base": 0.650527210759867,
298
+ "fve_aligned": 0.6851702085964343,
299
+ "fve_base_by_layer": [
300
+ 0.6429182481391268,
301
+ 0.6468658194492001,
302
+ 0.661797509455556
303
+ ],
304
+ "fve_aligned_by_layer": [
305
+ 0.6799573065098667,
306
+ 0.6851004518139425,
307
+ 0.6904528172228349
308
+ ],
309
+ "val_fve_base_by_layer": [
310
+ 0.6429182481391268,
311
+ 0.6468658194492001,
312
+ 0.661797509455556
313
+ ],
314
+ "val_fve_aligned_by_layer": [
315
+ 0.6799573065098667,
316
+ 0.6851004518139425,
317
+ 0.6904528172228349
318
+ ],
319
+ "dead_neuron_fraction": 0.9710712081775008,
320
+ "l0_sparsity_base": 136.68414228253383,
321
+ "l0_sparsity_aligned": 134.29049186461583,
322
+ "l0_base_by_layer": [
323
+ 125.3520869819031,
324
+ 134.57740440747227,
325
+ 150.12292031523643
326
+ ],
327
+ "l0_aligned_by_layer": [
328
+ 126.03602962638645,
329
+ 131.27096103327494,
330
+ 145.56447022767074
331
+ ],
332
+ "val_l0_base": 137.48267991630195,
333
+ "val_l0_aligned": 134.74987800957643,
334
+ "val_l0_base_by_layer": [
335
+ 126.13634383116717,
336
+ 135.3324607329843,
337
+ 150.9792212041885
338
+ ],
339
+ "val_l0_aligned_by_layer": [
340
+ 126.54193938090539,
341
+ 131.6101112565445,
342
+ 146.0975676531567
343
+ ],
344
+ "superposition_fraction": 0.0
345
+ }
v1/assembled_activations/llama32-3b-kto/metrics/training_metrics.json ADDED
@@ -0,0 +1,524 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epochs": [
3
+ 1,
4
+ 2,
5
+ 3,
6
+ 4,
7
+ 5,
8
+ 6,
9
+ 7,
10
+ 8
11
+ ],
12
+ "train_loss": [
13
+ 0.5933123942327583,
14
+ 0.442794848935895,
15
+ 0.42607008457462203,
16
+ 0.38806798507856893,
17
+ 0.3780940956040843,
18
+ 0.37194959002493144,
19
+ 0.36910743747826424,
20
+ 0.36662243225737334
21
+ ],
22
+ "val_loss": [
23
+ 0.4522804942430626,
24
+ 0.43734881515902374,
25
+ 0.41274881846617656,
26
+ 0.38737851022425746,
27
+ 0.38028734547929616,
28
+ 0.3786773179214038,
29
+ 0.3776190936253333,
30
+ 0.37696840079667054
31
+ ],
32
+ "train_fve_base": [
33
+ 0.4814383069961313,
34
+ 0.6003544950304293,
35
+ 0.6189264595265007,
36
+ 0.6378410120352763,
37
+ 0.6485570505287518,
38
+ 0.6568282935540021,
39
+ 0.6628907306529733,
40
+ 0.6666664291875654
41
+ ],
42
+ "train_fve_aligned": [
43
+ 0.4897083103253359,
44
+ 0.5930533013842082,
45
+ 0.6123184734662931,
46
+ 0.6680812882667247,
47
+ 0.680344003651202,
48
+ 0.6861663695252993,
49
+ 0.6893872699748683,
50
+ 0.6921899963803133
51
+ ],
52
+ "val_fve_base": [
53
+ 0.5845968798193008,
54
+ 0.6098661441453465,
55
+ 0.6222324986108311,
56
+ 0.635092594860736,
57
+ 0.6419743776945543,
58
+ 0.6467851612580384,
59
+ 0.6494424973482861,
60
+ 0.650527210759867
61
+ ],
62
+ "val_fve_aligned": [
63
+ 0.5853793015654799,
64
+ 0.5955420280002175,
65
+ 0.6358089334677651,
66
+ 0.671886538023724,
67
+ 0.680234924036795,
68
+ 0.6819715406258069,
69
+ 0.6842551374934731,
70
+ 0.6851702085964343
71
+ ],
72
+ "val_fve_base_by_layer": [
73
+ [
74
+ 0.5732664069580158,
75
+ 0.5763397326020046,
76
+ 0.6041844409173696
77
+ ],
78
+ [
79
+ 0.5986687528525347,
80
+ 0.6034028695515937,
81
+ 0.6275267547961929
82
+ ],
83
+ [
84
+ 0.6165900526870608,
85
+ 0.6112354472669631,
86
+ 0.6388719418910161
87
+ ],
88
+ [
89
+ 0.6276468862413735,
90
+ 0.6292668268942708,
91
+ 0.6483640190194414
92
+ ],
93
+ [
94
+ 0.6339579698927116,
95
+ 0.6365792960396612,
96
+ 0.6553858159724331
97
+ ],
98
+ [
99
+ 0.6386778891398645,
100
+ 0.6424399002684349,
101
+ 0.6592376335129064
102
+ ],
103
+ [
104
+ 0.6416899889551532,
105
+ 0.6455407226897035,
106
+ 0.6610967317176739
107
+ ],
108
+ [
109
+ 0.6429182481391268,
110
+ 0.6468658194492001,
111
+ 0.661797509455556
112
+ ]
113
+ ],
114
+ "val_fve_aligned_by_layer": [
115
+ [
116
+ 0.5565477425515339,
117
+ 0.5519646746325867,
118
+ 0.647625432276601
119
+ ],
120
+ [
121
+ 0.5665813191398901,
122
+ 0.5605105738989345,
123
+ 0.6595341338537126
124
+ ],
125
+ [
126
+ 0.6707772475262587,
127
+ 0.5713420142053933,
128
+ 0.6653074703291448
129
+ ],
130
+ [
131
+ 0.6783266744688543,
132
+ 0.6637531731141175,
133
+ 0.6735797000180989
134
+ ],
135
+ [
136
+ 0.6752785547241491,
137
+ 0.6786753744355047,
138
+ 0.6867507864667483
139
+ ],
140
+ [
141
+ 0.6770623098493247,
142
+ 0.68111452007793,
143
+ 0.687737735154117
144
+ ],
145
+ [
146
+ 0.6787484924830691,
147
+ 0.6842160321655074,
148
+ 0.689800825106536
149
+ ],
150
+ [
151
+ 0.6799573065098667,
152
+ 0.6851004518139425,
153
+ 0.6904528172228349
154
+ ]
155
+ ],
156
+ "train_fve_base_by_layer": [
157
+ [
158
+ 0.4730913452216915,
159
+ 0.47967155965809505,
160
+ 0.4915519679629587
161
+ ],
162
+ [
163
+ 0.5891723040593554,
164
+ 0.5918736373271708,
165
+ 0.6200174819080021
166
+ ],
167
+ [
168
+ 0.6081104554840506,
169
+ 0.6112697695474909,
170
+ 0.6373990934909747
171
+ ],
172
+ [
173
+ 0.631499816713038,
174
+ 0.6302260103381592,
175
+ 0.6517971492760096
176
+ ],
177
+ [
178
+ 0.6413867825732479,
179
+ 0.6431660363010266,
180
+ 0.6611182747427232
181
+ ],
182
+ [
183
+ 0.6489884858234404,
184
+ 0.651991655585784,
185
+ 0.6695046753682939
186
+ ],
187
+ [
188
+ 0.6551149910956325,
189
+ 0.658601807177032,
190
+ 0.6749553329333602
191
+ ],
192
+ [
193
+ 0.6590984342951421,
194
+ 0.6626518236846166,
195
+ 0.6782489653504946
196
+ ]
197
+ ],
198
+ "train_fve_aligned_by_layer": [
199
+ [
200
+ 0.48375090509566104,
201
+ 0.4794864978138912,
202
+ 0.5058874788857731
203
+ ],
204
+ [
205
+ 0.5638634942110978,
206
+ 0.5601495567359078,
207
+ 0.6551467945752475
208
+ ],
209
+ [
210
+ 0.6013194078669796,
211
+ 0.5714226390490111,
212
+ 0.6642133136694708
213
+ ],
214
+ [
215
+ 0.6761792779177583,
216
+ 0.6555343266474317,
217
+ 0.6725302028224542
218
+ ],
219
+ [
220
+ 0.6819632140015971,
221
+ 0.6775305208516831,
222
+ 0.6815382182702501
223
+ ],
224
+ [
225
+ 0.6810176338080558,
226
+ 0.6852003142141976,
227
+ 0.6922811020276592
228
+ ],
229
+ [
230
+ 0.6838987355544006,
231
+ 0.6887385682341514,
232
+ 0.6955244492454551
233
+ ],
234
+ [
235
+ 0.6868440483093818,
236
+ 0.6917720467321927,
237
+ 0.6979538341467657
238
+ ]
239
+ ],
240
+ "dead_neurons": [
241
+ 0.9377401455827993,
242
+ 0.9667631407891946,
243
+ 0.9680073846172883,
244
+ 0.9687811607376429,
245
+ 0.9695221584092367,
246
+ 0.9703716389004139,
247
+ 0.9709274034296916,
248
+ 0.9710712081775008
249
+ ],
250
+ "l0_base": [
251
+ 168.06772301757954,
252
+ 126.81064258760273,
253
+ 123.2507261019037,
254
+ 126.34850245808136,
255
+ 130.52682192257393,
256
+ 132.99421610834997,
257
+ 135.4419259128693,
258
+ 136.68414228253383
259
+ ],
260
+ "l0_aligned": [
261
+ 159.23378112311957,
262
+ 122.26635424630098,
263
+ 119.10803059201038,
264
+ 123.13043885862903,
265
+ 126.86929286994645,
266
+ 129.55245873695637,
267
+ 132.93504854593093,
268
+ 134.29049186461583
269
+ ],
270
+ "l0_base_by_layer": [
271
+ [
272
+ 161.7963915645067,
273
+ 167.02606903093988,
274
+ 175.3806917688266
275
+ ],
276
+ [
277
+ 117.48657326328079,
278
+ 123.98622664915354,
279
+ 138.95911777583188
280
+ ],
281
+ [
282
+ 115.22205195563339,
283
+ 121.02900612959719,
284
+ 133.501112813777
285
+ ],
286
+ [
287
+ 115.18999927028605,
288
+ 123.78610259778166,
289
+ 140.06939579684763
290
+ ],
291
+ [
292
+ 117.65137915936953,
293
+ 129.656158785756,
294
+ 144.27291301809692
295
+ ],
296
+ [
297
+ 120.36387186223,
298
+ 131.73347197898423,
299
+ 146.88528896672506
300
+ ],
301
+ [
302
+ 123.50923088149446,
303
+ 133.5356830122592,
304
+ 149.28084865732632
305
+ ],
306
+ [
307
+ 125.3520869819031,
308
+ 134.57740440747227,
309
+ 150.12292031523643
310
+ ]
311
+ ],
312
+ "l0_aligned_by_layer": [
313
+ [
314
+ 155.430823117338,
315
+ 157.74321366024517,
316
+ 164.5272913018097
317
+ ],
318
+ [
319
+ 115.72462419731465,
320
+ 117.97945855224752,
321
+ 133.09497227086982
322
+ ],
323
+ [
324
+ 113.77422650321074,
325
+ 113.63875510799767,
326
+ 129.91110259778168
327
+ ],
328
+ [
329
+ 114.65721687098657,
330
+ 118.46561223000583,
331
+ 136.26848000583772
332
+ ],
333
+ [
334
+ 117.2795716579101,
335
+ 124.29976649153532,
336
+ 139.0285318155283
337
+ ],
338
+ [
339
+ 120.08021380618797,
340
+ 126.71048598949211,
341
+ 141.86666301809692
342
+ ],
343
+ [
344
+ 124.03427831290135,
345
+ 129.95809617629888,
346
+ 144.81275539988326
347
+ ],
348
+ [
349
+ 126.03602962638645,
350
+ 131.27096103327494,
351
+ 145.56447022767074
352
+ ]
353
+ ],
354
+ "val_l0_base": [
355
+ 133.6790515040852,
356
+ 126.03499806988302,
357
+ 122.7813797396515,
358
+ 129.9280155042079,
359
+ 132.89138482378416,
360
+ 134.8359382988895,
361
+ 136.81023291143447,
362
+ 137.48267991630195
363
+ ],
364
+ "val_l0_aligned": [
365
+ 127.6454003219205,
366
+ 121.64332736349854,
367
+ 117.30543074682745,
368
+ 127.08719019864866,
369
+ 128.0597404360147,
370
+ 131.95315767457973,
371
+ 134.10319015742596,
372
+ 134.74987800957643
373
+ ],
374
+ "val_l0_base_by_layer": [
375
+ [
376
+ 125.309500449615,
377
+ 128.9821662303665,
378
+ 146.745473412319
379
+ ],
380
+ [
381
+ 115.72027707724047,
382
+ 125.55879147514622,
383
+ 136.8259162303665
384
+ ],
385
+ [
386
+ 113.56468152750224,
387
+ 120.74323736060977,
388
+ 134.03621294111483
389
+ ],
390
+ [
391
+ 117.10433029254693,
392
+ 127.62696335078535,
393
+ 145.05273781022476
394
+ ],
395
+ [
396
+ 120.67233859556508,
397
+ 130.8184446165075,
398
+ 147.18335520160136
399
+ ],
400
+ [
401
+ 122.73429319371728,
402
+ 132.8148451400677,
403
+ 148.95866058509387
404
+ ],
405
+ [
406
+ 125.38885254385583,
407
+ 134.5613547120419,
408
+ 150.48047562044954
409
+ ],
410
+ [
411
+ 126.13634383116717,
412
+ 135.3324607329843,
413
+ 150.9792212041885
414
+ ]
415
+ ],
416
+ "val_l0_aligned_by_layer": [
417
+ [
418
+ 124.38279887893437,
419
+ 120.35596641820139,
420
+ 138.19742588222962
421
+ ],
422
+ [
423
+ 114.432973412319,
424
+ 117.49503709882966,
425
+ 133.00196335078533
426
+ ],
427
+ [
428
+ 109.7994109947644,
429
+ 110.60089442867259,
430
+ 131.5159795471511
431
+ ],
432
+ [
433
+ 116.04553885734518,
434
+ 124.60585732984293,
435
+ 140.61016582069595
436
+ ],
437
+ [
438
+ 119.92277486910994,
439
+ 124.03735822907294,
440
+ 140.2190772251309
441
+ ],
442
+ [
443
+ 122.63519854320906,
444
+ 129.07717060668307,
445
+ 144.1470876963351
446
+ ],
447
+ [
448
+ 125.78043193717278,
449
+ 130.91977535866943,
450
+ 145.60934775786873
451
+ ],
452
+ [
453
+ 126.54193938090539,
454
+ 131.6101112565445,
455
+ 146.0975676531567
456
+ ]
457
+ ],
458
+ "self_recon": [
459
+ 0.3596775165024516,
460
+ 0.2799183193418065,
461
+ 0.26739436908208064,
462
+ 0.24186065978323307,
463
+ 0.23369103840618055,
464
+ 0.22850491952910038,
465
+ 0.22538104173180937,
466
+ 0.22313513932559342
467
+ ],
468
+ "cross_recon": [
469
+ 0.38041265005266467,
470
+ 0.3039850652113479,
471
+ 0.29148193161284375,
472
+ 0.26661263893413323,
473
+ 0.2592871804127412,
474
+ 0.25357689010387124,
475
+ 0.2501190767187997,
476
+ 0.24781497845820913
477
+ ],
478
+ "sparsity": [
479
+ 0.0814698132715132,
480
+ 0.04128250095208207,
481
+ 0.04208294076924427,
482
+ 0.03956226723554603,
483
+ 0.04068818295856212,
484
+ 0.04201391230131267,
485
+ 0.04367876260541923,
486
+ 0.04436129915717744
487
+ ],
488
+ "val_self_recon": [
489
+ 0.28703898846791054,
490
+ 0.27524666722220276,
491
+ 0.25847701931186995,
492
+ 0.2406287566997618,
493
+ 0.23478099087458007,
494
+ 0.2326575024589818,
495
+ 0.23096176958521,
496
+ 0.23029655413165767
497
+ ],
498
+ "val_cross_recon": [
499
+ 0.311376446516726,
500
+ 0.29898168794147634,
501
+ 0.2829330632979957,
502
+ 0.26673236247444654,
503
+ 0.2611719256764307,
504
+ 0.25862815227183994,
505
+ 0.2567455348856162,
506
+ 0.2561143940187874
507
+ ],
508
+ "val_sparsity": [
509
+ 0.04069092366551853,
510
+ 0.042509469140262504,
511
+ 0.04109857126056212,
512
+ 0.04005680582364192,
513
+ 0.04103758367490394,
514
+ 0.04256855082059406,
515
+ 0.04395910826419037,
516
+ 0.044226086713100604
517
+ ],
518
+ "layers": [
519
+ 23,
520
+ 24,
521
+ 25
522
+ ],
523
+ "topk_mode": "model_balanced_layer_agg"
524
+ }
v1/assembled_activations/llama32-3b-kto/plots/aligned_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: 98a2d17b0b6336e24bc837aeee5e376c6e3ed95f22ee06b197bf281babfd7122
  • Pointer size: 131 Bytes
  • Size of remote file: 193 kB
v1/assembled_activations/llama32-3b-kto/plots/base_decoder_norm_heatmap.png ADDED

Git LFS Details

  • SHA256: 14d3548a4f969966bef35766a40644d32987925d28f5100bc5fa91108e345d9a
  • Pointer size: 131 Bytes
  • Size of remote file: 192 kB
v1/assembled_activations/llama32-3b-kto/plots/cf_shift_by_layer.png ADDED

Git LFS Details

  • SHA256: c0cfb9ddb61e9e4dea4628970ea95599322e1e2d0bd64921890acbb470e36c19
  • Pointer size: 130 Bytes
  • Size of remote file: 44.6 kB
v1/assembled_activations/llama32-3b-kto/plots/cf_shift_p95_by_layer.png ADDED

Git LFS Details

  • SHA256: 2963836862349691f6788b051eae48084d3f7c4001dca7c19b2b35b6585061fd
  • Pointer size: 130 Bytes
  • Size of remote file: 69.7 kB
v1/assembled_activations/llama32-3b-kto/plots/class_distribution_multilayer.png ADDED

Git LFS Details

  • SHA256: fd3ff07c4a0aeaae248f72f7bd0a65dc6c04bf1aee67db63a68c379808634d3e
  • Pointer size: 130 Bytes
  • Size of remote file: 60.8 kB
v1/assembled_activations/llama32-3b-kto/plots/class_distribution_primary.png ADDED

Git LFS Details

  • SHA256: ab54c713eb4146a24a787239970a4faad510a00eb3353ad36faea95692749cef
  • Pointer size: 131 Bytes
  • Size of remote file: 144 kB
v1/assembled_activations/llama32-3b-kto/plots/cross_layer_cosine_drift_by_stream.png ADDED

Git LFS Details

  • SHA256: 067d783ab2932acf4c4b646e40cfebdb1bcb2f16709306d034c75e7b59868862
  • Pointer size: 130 Bytes
  • Size of remote file: 51.3 kB
v1/assembled_activations/llama32-3b-kto/plots/decoder_norm_ratio_by_layer.png ADDED

Git LFS Details

  • SHA256: 08759185b138eb5aeaed7e7e0bd9769f9a93af15d9daf1efded918465861c1d6
  • Pointer size: 130 Bytes
  • Size of remote file: 45.2 kB
v1/assembled_activations/llama32-3b-kto/plots/feature_layer_trajectories.png ADDED

Git LFS Details

  • SHA256: 75ab8d20cb8b859e91e5c3b8f2b41f777118536cbc7ae63587200f28e6230a5a
  • Pointer size: 131 Bytes
  • Size of remote file: 548 kB
v1/assembled_activations/llama32-3b-kto/plots/feature_sharing_ratio_by_layer.png ADDED

Git LFS Details

  • SHA256: be836db7d8451e4793d1e064ded56d2617c885dbf770876e81bd6a63c8e76329
  • Pointer size: 130 Bytes
  • Size of remote file: 47.7 kB
v1/assembled_activations/llama32-3b-kto/plots/fve_by_layer.png ADDED

Git LFS Details

  • SHA256: 2cfb5f320004421c379af1901caed5fa97d5db15eb818741e5d3787dea87d1fb
  • Pointer size: 130 Bytes
  • Size of remote file: 83.2 kB
v1/assembled_activations/llama32-3b-kto/plots/l0_by_layer.png ADDED

Git LFS Details

  • SHA256: 5b440fb21cc24600e4b945b142807b139e416b1be6ec6954dbe4c26ec93a4b3d
  • Pointer size: 130 Bytes
  • Size of remote file: 89.9 kB
v1/assembled_activations/llama32-3b-kto/plots/layer_concentration_entropy.png ADDED

Git LFS Details

  • SHA256: 0304978c2d1033aed5db64a6a6daf5dc02d2a1ca96919e0a6d5a8a5910c68848
  • Pointer size: 130 Bytes
  • Size of remote file: 91.9 kB
v1/assembled_activations/llama32-3b-kto/plots/loss_curves.png ADDED

Git LFS Details

  • SHA256: 4d96c2627a7f500d0b95dccb3bd5fb80ea8d3212023b98d0103e3df254005389
  • Pointer size: 131 Bytes
  • Size of remote file: 324 kB
v1/assembled_activations/llama32-3b-kto/plots/max_norm_layer_migration.png ADDED

Git LFS Details

  • SHA256: 06f0d0213e47c3438dd4d0e7c9473f1341d51d70ef96048fda74ffa2b6a9a7f2
  • Pointer size: 130 Bytes
  • Size of remote file: 81.4 kB
v1/assembled_activations/llama32-3b-kto/plots/rho_histogram_by_layer.png ADDED

Git LFS Details

  • SHA256: 0a0f955803823924ac7fa35fdafb9032eeec82d8786d8858f933fcd8d62aa03d
  • Pointer size: 130 Bytes
  • Size of remote file: 47 kB
v1/assembled_activations/llama32-3b-kto/plots/rho_theta_scatter_by_layer.png ADDED

Git LFS Details

  • SHA256: dba480ac7d3b855aacc12cf07b2227f121714c56c87d59880e17ecb695b90d07
  • Pointer size: 131 Bytes
  • Size of remote file: 211 kB
v1/assembled_activations/llama32-3b-kto/plots/superposition_by_layer.png ADDED

Git LFS Details

  • SHA256: 7f666e4b8d3181ba9fc51042d458d649f75dfac5b938aab4262a327746e314a8
  • Pointer size: 130 Bytes
  • Size of remote file: 97.9 kB
v1/assembled_activations/llama32-3b-kto/plots/theta_by_layer.png ADDED

Git LFS Details

  • SHA256: 33f0b20685068304693fe8b16dd9a80b8e33d7507a37a55811a1a1d33ba90abb
  • Pointer size: 130 Bytes
  • Size of remote file: 49 kB