Add assembled activations for llama32-3b-kto
Browse files- .gitattributes +2 -0
- v1/assembled_activations/llama32-3b-kto/activations/activations.pt +3 -0
- v1/assembled_activations/llama32-3b-kto/checkpoints/final.pt +3 -0
- v1/assembled_activations/llama32-3b-kto/features/counterfactual_scores.csv +0 -0
- v1/assembled_activations/llama32-3b-kto/features/counterfactual_scores_by_layer.csv +0 -0
- v1/assembled_activations/llama32-3b-kto/features/cross_layer_cosine_drift.csv +3 -0
- v1/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv +3 -0
- v1/assembled_activations/llama32-3b-kto/features/feature_activations.pt +3 -0
- v1/assembled_activations/llama32-3b-kto/features/feature_classification.csv +0 -0
- v1/assembled_activations/llama32-3b-kto/features/merged_classification.csv +0 -0
- v1/assembled_activations/llama32-3b-kto/features/model_layer_stream_patterns.csv +0 -0
- v1/assembled_activations/llama32-3b-kto/features/superposition_analysis.json +271 -0
- v1/assembled_activations/llama32-3b-kto/metrics/aggregate_metrics.json +345 -0
- v1/assembled_activations/llama32-3b-kto/metrics/training_metrics.json +524 -0
- v1/assembled_activations/llama32-3b-kto/plots/aligned_decoder_norm_heatmap.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/base_decoder_norm_heatmap.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/cf_shift_by_layer.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/cf_shift_p95_by_layer.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/class_distribution_multilayer.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/class_distribution_primary.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/cross_layer_cosine_drift_by_stream.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/decoder_norm_ratio_by_layer.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/feature_layer_trajectories.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/feature_sharing_ratio_by_layer.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/fve_by_layer.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/l0_by_layer.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/layer_concentration_entropy.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/loss_curves.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/max_norm_layer_migration.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/rho_histogram_by_layer.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/rho_theta_scatter_by_layer.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/superposition_by_layer.png +3 -0
- v1/assembled_activations/llama32-3b-kto/plots/theta_by_layer.png +3 -0
.gitattributes
CHANGED
|
@@ -76,3 +76,5 @@ v1/assembled_activations/llama32-3b-dpo/features/cross_layer_cosine_drift.csv fi
|
|
| 76 |
v1/assembled_activations/llama32-3b-dpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
|
| 77 |
v1/assembled_activations/llama32-3b-grpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
|
| 78 |
v1/assembled_activations/llama32-3b-grpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 76 |
v1/assembled_activations/llama32-3b-dpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
|
| 77 |
v1/assembled_activations/llama32-3b-grpo/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
|
| 78 |
v1/assembled_activations/llama32-3b-grpo/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
|
| 79 |
+
v1/assembled_activations/llama32-3b-kto/features/cross_layer_cosine_drift.csv filter=lfs diff=lfs merge=lfs -text
|
| 80 |
+
v1/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv filter=lfs diff=lfs merge=lfs -text
|
v1/assembled_activations/llama32-3b-kto/activations/activations.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:031a8ce6e30f0e94cf173814bde688d7e1879b77987dec414f459a448d62ef6f
|
| 3 |
+
size 4496133925
|
v1/assembled_activations/llama32-3b-kto/checkpoints/final.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:40324b29037cf39bebfb2753779f737db78f0948c0ba9adab6364b25e1cca1ca
|
| 3 |
+
size 10873439922
|
v1/assembled_activations/llama32-3b-kto/features/counterfactual_scores.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
v1/assembled_activations/llama32-3b-kto/features/counterfactual_scores_by_layer.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
v1/assembled_activations/llama32-3b-kto/features/cross_layer_cosine_drift.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:67419b2031b8730c4ba88558182d2d40e821df0cadd7fcaca4309cadbec2e92f
|
| 3 |
+
size 26853060
|
v1/assembled_activations/llama32-3b-kto/features/decoder_layer_profiles.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bb7314260fbc192b114f5eede5cf464ace89bafec2a823d2f15081e60fc7a9f1
|
| 3 |
+
size 13167539
|
v1/assembled_activations/llama32-3b-kto/features/feature_activations.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:366df02792e05c657cc565e2810ae2860e7c5f4994f7d51b32a5dc92ec5e128f
|
| 3 |
+
size 35935153173
|
v1/assembled_activations/llama32-3b-kto/features/feature_classification.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
v1/assembled_activations/llama32-3b-kto/features/merged_classification.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
v1/assembled_activations/llama32-3b-kto/features/model_layer_stream_patterns.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
v1/assembled_activations/llama32-3b-kto/features/superposition_analysis.json
ADDED
|
@@ -0,0 +1,271 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"analysis_kind": "multilayer_decoder_cosine_candidate_matches",
|
| 3 |
+
"description": "Candidate cross-layer superposition screen: aligned decoder vectors are matched against base decoder vectors across all selected layers/features. This is a geometry-derived screen, not a causal decomposition.",
|
| 4 |
+
"layers": [
|
| 5 |
+
23,
|
| 6 |
+
24,
|
| 7 |
+
25
|
| 8 |
+
],
|
| 9 |
+
"top_k": 5,
|
| 10 |
+
"cosine_threshold": 0.5,
|
| 11 |
+
"regression_candidates": 256,
|
| 12 |
+
"lasso_alpha": 0.01,
|
| 13 |
+
"n_features_analyzed": 5,
|
| 14 |
+
"features": {
|
| 15 |
+
"2344": {
|
| 16 |
+
"feature_id": 2344,
|
| 17 |
+
"primary_class": "aligned_only",
|
| 18 |
+
"target_layer": 25,
|
| 19 |
+
"r2": -4.07232970456306e-07,
|
| 20 |
+
"n_nonzero": 0,
|
| 21 |
+
"is_superposition": false,
|
| 22 |
+
"constituent_features": [],
|
| 23 |
+
"top_base_matches": [
|
| 24 |
+
{
|
| 25 |
+
"base_feature_id": 24296,
|
| 26 |
+
"base_layer": 24,
|
| 27 |
+
"cosine": 0.45633772015571594,
|
| 28 |
+
"abs_cosine": 0.45633772015571594,
|
| 29 |
+
"is_same_feature": false,
|
| 30 |
+
"is_cross_layer": true
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"base_feature_id": 18687,
|
| 34 |
+
"base_layer": 25,
|
| 35 |
+
"cosine": 0.4250088334083557,
|
| 36 |
+
"abs_cosine": 0.4250088334083557,
|
| 37 |
+
"is_same_feature": false,
|
| 38 |
+
"is_cross_layer": false
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"base_feature_id": 2865,
|
| 42 |
+
"base_layer": 24,
|
| 43 |
+
"cosine": -0.4132736921310425,
|
| 44 |
+
"abs_cosine": 0.4132736921310425,
|
| 45 |
+
"is_same_feature": false,
|
| 46 |
+
"is_cross_layer": true
|
| 47 |
+
},
|
| 48 |
+
{
|
| 49 |
+
"base_feature_id": 7175,
|
| 50 |
+
"base_layer": 23,
|
| 51 |
+
"cosine": -0.39883679151535034,
|
| 52 |
+
"abs_cosine": 0.39883679151535034,
|
| 53 |
+
"is_same_feature": false,
|
| 54 |
+
"is_cross_layer": true
|
| 55 |
+
},
|
| 56 |
+
{
|
| 57 |
+
"base_feature_id": 6201,
|
| 58 |
+
"base_layer": 23,
|
| 59 |
+
"cosine": -0.39130496978759766,
|
| 60 |
+
"abs_cosine": 0.39130496978759766,
|
| 61 |
+
"is_same_feature": false,
|
| 62 |
+
"is_cross_layer": true
|
| 63 |
+
}
|
| 64 |
+
]
|
| 65 |
+
},
|
| 66 |
+
"11037": {
|
| 67 |
+
"feature_id": 11037,
|
| 68 |
+
"primary_class": "aligned_only",
|
| 69 |
+
"target_layer": 25,
|
| 70 |
+
"r2": -0.0009124147442902686,
|
| 71 |
+
"n_nonzero": 0,
|
| 72 |
+
"is_superposition": false,
|
| 73 |
+
"constituent_features": [],
|
| 74 |
+
"top_base_matches": [
|
| 75 |
+
{
|
| 76 |
+
"base_feature_id": 13612,
|
| 77 |
+
"base_layer": 25,
|
| 78 |
+
"cosine": -0.6919379234313965,
|
| 79 |
+
"abs_cosine": 0.6919379234313965,
|
| 80 |
+
"is_same_feature": false,
|
| 81 |
+
"is_cross_layer": false
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"base_feature_id": 9490,
|
| 85 |
+
"base_layer": 25,
|
| 86 |
+
"cosine": -0.6915594339370728,
|
| 87 |
+
"abs_cosine": 0.6915594339370728,
|
| 88 |
+
"is_same_feature": false,
|
| 89 |
+
"is_cross_layer": false
|
| 90 |
+
},
|
| 91 |
+
{
|
| 92 |
+
"base_feature_id": 13762,
|
| 93 |
+
"base_layer": 25,
|
| 94 |
+
"cosine": 0.6774390935897827,
|
| 95 |
+
"abs_cosine": 0.6774390935897827,
|
| 96 |
+
"is_same_feature": false,
|
| 97 |
+
"is_cross_layer": false
|
| 98 |
+
},
|
| 99 |
+
{
|
| 100 |
+
"base_feature_id": 2287,
|
| 101 |
+
"base_layer": 25,
|
| 102 |
+
"cosine": -0.6305816173553467,
|
| 103 |
+
"abs_cosine": 0.6305816173553467,
|
| 104 |
+
"is_same_feature": false,
|
| 105 |
+
"is_cross_layer": false
|
| 106 |
+
},
|
| 107 |
+
{
|
| 108 |
+
"base_feature_id": 15355,
|
| 109 |
+
"base_layer": 24,
|
| 110 |
+
"cosine": -0.6074807643890381,
|
| 111 |
+
"abs_cosine": 0.6074807643890381,
|
| 112 |
+
"is_same_feature": false,
|
| 113 |
+
"is_cross_layer": true
|
| 114 |
+
}
|
| 115 |
+
]
|
| 116 |
+
},
|
| 117 |
+
"18301": {
|
| 118 |
+
"feature_id": 18301,
|
| 119 |
+
"primary_class": "aligned_only",
|
| 120 |
+
"target_layer": 23,
|
| 121 |
+
"r2": -4.195340993251584e-05,
|
| 122 |
+
"n_nonzero": 0,
|
| 123 |
+
"is_superposition": false,
|
| 124 |
+
"constituent_features": [],
|
| 125 |
+
"top_base_matches": [
|
| 126 |
+
{
|
| 127 |
+
"base_feature_id": 12929,
|
| 128 |
+
"base_layer": 23,
|
| 129 |
+
"cosine": -0.5700091123580933,
|
| 130 |
+
"abs_cosine": 0.5700091123580933,
|
| 131 |
+
"is_same_feature": false,
|
| 132 |
+
"is_cross_layer": false
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"base_feature_id": 17075,
|
| 136 |
+
"base_layer": 23,
|
| 137 |
+
"cosine": -0.53563392162323,
|
| 138 |
+
"abs_cosine": 0.53563392162323,
|
| 139 |
+
"is_same_feature": false,
|
| 140 |
+
"is_cross_layer": false
|
| 141 |
+
},
|
| 142 |
+
{
|
| 143 |
+
"base_feature_id": 15355,
|
| 144 |
+
"base_layer": 24,
|
| 145 |
+
"cosine": -0.5292654037475586,
|
| 146 |
+
"abs_cosine": 0.5292654037475586,
|
| 147 |
+
"is_same_feature": false,
|
| 148 |
+
"is_cross_layer": true
|
| 149 |
+
},
|
| 150 |
+
{
|
| 151 |
+
"base_feature_id": 6201,
|
| 152 |
+
"base_layer": 23,
|
| 153 |
+
"cosine": -0.5044533014297485,
|
| 154 |
+
"abs_cosine": 0.5044533014297485,
|
| 155 |
+
"is_same_feature": false,
|
| 156 |
+
"is_cross_layer": false
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"base_feature_id": 24296,
|
| 160 |
+
"base_layer": 24,
|
| 161 |
+
"cosine": 0.5009859800338745,
|
| 162 |
+
"abs_cosine": 0.5009859800338745,
|
| 163 |
+
"is_same_feature": false,
|
| 164 |
+
"is_cross_layer": true
|
| 165 |
+
}
|
| 166 |
+
]
|
| 167 |
+
},
|
| 168 |
+
"19481": {
|
| 169 |
+
"feature_id": 19481,
|
| 170 |
+
"primary_class": "aligned_only",
|
| 171 |
+
"target_layer": 24,
|
| 172 |
+
"r2": -0.00041131219288481624,
|
| 173 |
+
"n_nonzero": 0,
|
| 174 |
+
"is_superposition": false,
|
| 175 |
+
"constituent_features": [],
|
| 176 |
+
"top_base_matches": [
|
| 177 |
+
{
|
| 178 |
+
"base_feature_id": 15355,
|
| 179 |
+
"base_layer": 24,
|
| 180 |
+
"cosine": -0.607896089553833,
|
| 181 |
+
"abs_cosine": 0.607896089553833,
|
| 182 |
+
"is_same_feature": false,
|
| 183 |
+
"is_cross_layer": false
|
| 184 |
+
},
|
| 185 |
+
{
|
| 186 |
+
"base_feature_id": 3311,
|
| 187 |
+
"base_layer": 24,
|
| 188 |
+
"cosine": 0.5702545642852783,
|
| 189 |
+
"abs_cosine": 0.5702545642852783,
|
| 190 |
+
"is_same_feature": false,
|
| 191 |
+
"is_cross_layer": false
|
| 192 |
+
},
|
| 193 |
+
{
|
| 194 |
+
"base_feature_id": 20061,
|
| 195 |
+
"base_layer": 24,
|
| 196 |
+
"cosine": -0.5608834028244019,
|
| 197 |
+
"abs_cosine": 0.5608834028244019,
|
| 198 |
+
"is_same_feature": false,
|
| 199 |
+
"is_cross_layer": false
|
| 200 |
+
},
|
| 201 |
+
{
|
| 202 |
+
"base_feature_id": 16726,
|
| 203 |
+
"base_layer": 24,
|
| 204 |
+
"cosine": -0.5551276206970215,
|
| 205 |
+
"abs_cosine": 0.5551276206970215,
|
| 206 |
+
"is_same_feature": false,
|
| 207 |
+
"is_cross_layer": false
|
| 208 |
+
},
|
| 209 |
+
{
|
| 210 |
+
"base_feature_id": 9490,
|
| 211 |
+
"base_layer": 25,
|
| 212 |
+
"cosine": -0.5516860485076904,
|
| 213 |
+
"abs_cosine": 0.5516860485076904,
|
| 214 |
+
"is_same_feature": false,
|
| 215 |
+
"is_cross_layer": true
|
| 216 |
+
}
|
| 217 |
+
]
|
| 218 |
+
},
|
| 219 |
+
"24296": {
|
| 220 |
+
"feature_id": 24296,
|
| 221 |
+
"primary_class": "aligned_only",
|
| 222 |
+
"target_layer": 24,
|
| 223 |
+
"r2": -0.00016202182581004365,
|
| 224 |
+
"n_nonzero": 0,
|
| 225 |
+
"is_superposition": false,
|
| 226 |
+
"constituent_features": [],
|
| 227 |
+
"top_base_matches": [
|
| 228 |
+
{
|
| 229 |
+
"base_feature_id": 24296,
|
| 230 |
+
"base_layer": 24,
|
| 231 |
+
"cosine": 0.5111119151115417,
|
| 232 |
+
"abs_cosine": 0.5111119151115417,
|
| 233 |
+
"is_same_feature": true,
|
| 234 |
+
"is_cross_layer": false
|
| 235 |
+
},
|
| 236 |
+
{
|
| 237 |
+
"base_feature_id": 18109,
|
| 238 |
+
"base_layer": 23,
|
| 239 |
+
"cosine": -0.43007758259773254,
|
| 240 |
+
"abs_cosine": 0.43007758259773254,
|
| 241 |
+
"is_same_feature": false,
|
| 242 |
+
"is_cross_layer": true
|
| 243 |
+
},
|
| 244 |
+
{
|
| 245 |
+
"base_feature_id": 7175,
|
| 246 |
+
"base_layer": 23,
|
| 247 |
+
"cosine": -0.42721956968307495,
|
| 248 |
+
"abs_cosine": 0.42721956968307495,
|
| 249 |
+
"is_same_feature": false,
|
| 250 |
+
"is_cross_layer": true
|
| 251 |
+
},
|
| 252 |
+
{
|
| 253 |
+
"base_feature_id": 1879,
|
| 254 |
+
"base_layer": 23,
|
| 255 |
+
"cosine": -0.42496082186698914,
|
| 256 |
+
"abs_cosine": 0.42496082186698914,
|
| 257 |
+
"is_same_feature": false,
|
| 258 |
+
"is_cross_layer": true
|
| 259 |
+
},
|
| 260 |
+
{
|
| 261 |
+
"base_feature_id": 6747,
|
| 262 |
+
"base_layer": 23,
|
| 263 |
+
"cosine": -0.42284414172172546,
|
| 264 |
+
"abs_cosine": 0.42284414172172546,
|
| 265 |
+
"is_same_feature": false,
|
| 266 |
+
"is_cross_layer": true
|
| 267 |
+
}
|
| 268 |
+
]
|
| 269 |
+
}
|
| 270 |
+
}
|
| 271 |
+
}
|
v1/assembled_activations/llama32-3b-kto/metrics/aggregate_metrics.json
ADDED
|
@@ -0,0 +1,345 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"crosscoder_kind": "multilayer_sparc",
|
| 3 |
+
"layers": [
|
| 4 |
+
23,
|
| 5 |
+
24,
|
| 6 |
+
25
|
| 7 |
+
],
|
| 8 |
+
"topk_mode": "model_balanced_layer_agg",
|
| 9 |
+
"topk": 400,
|
| 10 |
+
"expansion_factor": 8,
|
| 11 |
+
"dict_size": 24576,
|
| 12 |
+
"forced_shared_fraction": 0.06,
|
| 13 |
+
"class_counts": {
|
| 14 |
+
"shared_redirected": 18354,
|
| 15 |
+
"shared_attenuated": 4083,
|
| 16 |
+
"shared_aligned": 1474,
|
| 17 |
+
"shared_intermediate": 660,
|
| 18 |
+
"aligned_only": 5
|
| 19 |
+
},
|
| 20 |
+
"multilayer_class_counts": {
|
| 21 |
+
"drifting_or_rotating": 22988,
|
| 22 |
+
"persistent_shared": 1388,
|
| 23 |
+
"mixed_or_ambiguous": 195,
|
| 24 |
+
"localized_aligned_only": 5
|
| 25 |
+
},
|
| 26 |
+
"classification_thresholds": {
|
| 27 |
+
"rho_base_only": 0.4,
|
| 28 |
+
"rho_aligned_only": 0.9799827247629397,
|
| 29 |
+
"rho_shared_low": 0.49517253515653925,
|
| 30 |
+
"rho_shared_high": 0.8529490783385397
|
| 31 |
+
},
|
| 32 |
+
"threshold_sensitivity": {
|
| 33 |
+
"original": {
|
| 34 |
+
"shared_redirected": 18354,
|
| 35 |
+
"shared_attenuated": 4083,
|
| 36 |
+
"shared_aligned": 1474,
|
| 37 |
+
"shared_intermediate": 660,
|
| 38 |
+
"aligned_only": 5
|
| 39 |
+
},
|
| 40 |
+
"perturbed": {
|
| 41 |
+
"delta_-0.05": {
|
| 42 |
+
"base_only": 0,
|
| 43 |
+
"aligned_only": 0,
|
| 44 |
+
"shared_aligned": 1474,
|
| 45 |
+
"shared_redirected": 22264,
|
| 46 |
+
"shared_intermediate": 829,
|
| 47 |
+
"shared_attenuated": 4,
|
| 48 |
+
"other": 5
|
| 49 |
+
},
|
| 50 |
+
"delta_+0.05": {
|
| 51 |
+
"base_only": 5,
|
| 52 |
+
"aligned_only": 5,
|
| 53 |
+
"shared_aligned": 0,
|
| 54 |
+
"shared_redirected": 919,
|
| 55 |
+
"shared_intermediate": 43,
|
| 56 |
+
"shared_attenuated": 23604,
|
| 57 |
+
"other": 0
|
| 58 |
+
}
|
| 59 |
+
},
|
| 60 |
+
"perturbation": 0.05
|
| 61 |
+
},
|
| 62 |
+
"class_counts_by_layer": {
|
| 63 |
+
"23": {
|
| 64 |
+
"aligned_only": 24452,
|
| 65 |
+
"base_only": 124
|
| 66 |
+
},
|
| 67 |
+
"24": {
|
| 68 |
+
"aligned_only": 21864,
|
| 69 |
+
"base_only": 2712
|
| 70 |
+
},
|
| 71 |
+
"25": {
|
| 72 |
+
"aligned_only": 24502,
|
| 73 |
+
"base_only": 74
|
| 74 |
+
}
|
| 75 |
+
},
|
| 76 |
+
"feature_sharing_ratio_by_layer": {
|
| 77 |
+
"23": 0.0,
|
| 78 |
+
"24": 0.0,
|
| 79 |
+
"25": 0.0
|
| 80 |
+
},
|
| 81 |
+
"decoder_amplification_by_layer": {
|
| 82 |
+
"23": {
|
| 83 |
+
"median": 1.047627542042361,
|
| 84 |
+
"p95": 1.362319406084207
|
| 85 |
+
},
|
| 86 |
+
"24": {
|
| 87 |
+
"median": 1.0430958416504477,
|
| 88 |
+
"p95": 1.3325615028387257
|
| 89 |
+
},
|
| 90 |
+
"25": {
|
| 91 |
+
"median": 1.0378214066487217,
|
| 92 |
+
"p95": 1.2347764635930154
|
| 93 |
+
}
|
| 94 |
+
},
|
| 95 |
+
"classification_thresholds_by_layer": {
|
| 96 |
+
"23": {
|
| 97 |
+
"rho_base_only": 0.3721634745107997,
|
| 98 |
+
"rho_aligned_only": 0.3721634745107997,
|
| 99 |
+
"rho_shared_low": 0.3721634745107997,
|
| 100 |
+
"rho_shared_high": 0.3721634745107997
|
| 101 |
+
},
|
| 102 |
+
"24": {
|
| 103 |
+
"rho_base_only": 0.46737811492284764,
|
| 104 |
+
"rho_aligned_only": 0.46737811492284764,
|
| 105 |
+
"rho_shared_low": 0.46737811492284764,
|
| 106 |
+
"rho_shared_high": 0.46737811492284764
|
| 107 |
+
},
|
| 108 |
+
"25": {
|
| 109 |
+
"rho_base_only": 0.40263814939318293,
|
| 110 |
+
"rho_aligned_only": 0.40263814939318293,
|
| 111 |
+
"rho_shared_low": 0.40263814939318293,
|
| 112 |
+
"rho_shared_high": 0.40263814939318293
|
| 113 |
+
}
|
| 114 |
+
},
|
| 115 |
+
"threshold_sensitivity_by_layer": {
|
| 116 |
+
"23": {
|
| 117 |
+
"original": {
|
| 118 |
+
"aligned_only": 24452,
|
| 119 |
+
"base_only": 124
|
| 120 |
+
},
|
| 121 |
+
"perturbed": {
|
| 122 |
+
"delta_-0.05": {
|
| 123 |
+
"base_only": 95,
|
| 124 |
+
"aligned_only": 24192,
|
| 125 |
+
"shared_aligned": 43,
|
| 126 |
+
"shared_redirected": 146,
|
| 127 |
+
"shared_intermediate": 100,
|
| 128 |
+
"shared_attenuated": 0,
|
| 129 |
+
"other": 0
|
| 130 |
+
},
|
| 131 |
+
"delta_+0.05": {
|
| 132 |
+
"base_only": 384,
|
| 133 |
+
"aligned_only": 24192,
|
| 134 |
+
"shared_aligned": 0,
|
| 135 |
+
"shared_redirected": 0,
|
| 136 |
+
"shared_intermediate": 0,
|
| 137 |
+
"shared_attenuated": 0,
|
| 138 |
+
"other": 0
|
| 139 |
+
}
|
| 140 |
+
},
|
| 141 |
+
"perturbation": 0.05
|
| 142 |
+
},
|
| 143 |
+
"24": {
|
| 144 |
+
"original": {
|
| 145 |
+
"aligned_only": 21864,
|
| 146 |
+
"base_only": 2712
|
| 147 |
+
},
|
| 148 |
+
"perturbed": {
|
| 149 |
+
"delta_-0.05": {
|
| 150 |
+
"base_only": 280,
|
| 151 |
+
"aligned_only": 10646,
|
| 152 |
+
"shared_aligned": 1374,
|
| 153 |
+
"shared_redirected": 11475,
|
| 154 |
+
"shared_intermediate": 801,
|
| 155 |
+
"shared_attenuated": 0,
|
| 156 |
+
"other": 0
|
| 157 |
+
},
|
| 158 |
+
"delta_+0.05": {
|
| 159 |
+
"base_only": 13930,
|
| 160 |
+
"aligned_only": 10646,
|
| 161 |
+
"shared_aligned": 0,
|
| 162 |
+
"shared_redirected": 0,
|
| 163 |
+
"shared_intermediate": 0,
|
| 164 |
+
"shared_attenuated": 0,
|
| 165 |
+
"other": 0
|
| 166 |
+
}
|
| 167 |
+
},
|
| 168 |
+
"perturbation": 0.05
|
| 169 |
+
},
|
| 170 |
+
"25": {
|
| 171 |
+
"original": {
|
| 172 |
+
"aligned_only": 24502,
|
| 173 |
+
"base_only": 74
|
| 174 |
+
},
|
| 175 |
+
"perturbed": {
|
| 176 |
+
"delta_-0.05": {
|
| 177 |
+
"base_only": 63,
|
| 178 |
+
"aligned_only": 24191,
|
| 179 |
+
"shared_aligned": 22,
|
| 180 |
+
"shared_redirected": 241,
|
| 181 |
+
"shared_intermediate": 59,
|
| 182 |
+
"shared_attenuated": 0,
|
| 183 |
+
"other": 0
|
| 184 |
+
},
|
| 185 |
+
"delta_+0.05": {
|
| 186 |
+
"base_only": 385,
|
| 187 |
+
"aligned_only": 24191,
|
| 188 |
+
"shared_aligned": 0,
|
| 189 |
+
"shared_redirected": 0,
|
| 190 |
+
"shared_intermediate": 0,
|
| 191 |
+
"shared_attenuated": 0,
|
| 192 |
+
"other": 0
|
| 193 |
+
}
|
| 194 |
+
},
|
| 195 |
+
"perturbation": 0.05
|
| 196 |
+
}
|
| 197 |
+
},
|
| 198 |
+
"counterfactual_shift_by_layer": {
|
| 199 |
+
"23": {
|
| 200 |
+
"aligned_only": {
|
| 201 |
+
"mean_shift": 0.05192818641662598,
|
| 202 |
+
"median_shift": 0.0,
|
| 203 |
+
"p95_abs_shift": 2.235345268249511,
|
| 204 |
+
"count": 5
|
| 205 |
+
},
|
| 206 |
+
"shared_aligned": {
|
| 207 |
+
"mean_shift": 0.17152424341133637,
|
| 208 |
+
"median_shift": -7.850458587199682e-06,
|
| 209 |
+
"p95_abs_shift": 4.667830324172954,
|
| 210 |
+
"count": 1474
|
| 211 |
+
},
|
| 212 |
+
"shared_attenuated": {
|
| 213 |
+
"mean_shift": -1.008251538443044e-06,
|
| 214 |
+
"median_shift": 0.0,
|
| 215 |
+
"p95_abs_shift": 0.0,
|
| 216 |
+
"count": 4083
|
| 217 |
+
},
|
| 218 |
+
"shared_intermediate": {
|
| 219 |
+
"mean_shift": 6.39913474182749e-07,
|
| 220 |
+
"median_shift": 0.0,
|
| 221 |
+
"p95_abs_shift": 0.0,
|
| 222 |
+
"count": 660
|
| 223 |
+
},
|
| 224 |
+
"shared_redirected": {
|
| 225 |
+
"mean_shift": 5.061595920287719e-07,
|
| 226 |
+
"median_shift": 0.0,
|
| 227 |
+
"p95_abs_shift": 0.0,
|
| 228 |
+
"count": 18354
|
| 229 |
+
}
|
| 230 |
+
},
|
| 231 |
+
"24": {
|
| 232 |
+
"aligned_only": {
|
| 233 |
+
"mean_shift": 0.4149253680070615,
|
| 234 |
+
"median_shift": 0.0,
|
| 235 |
+
"p95_abs_shift": 9.360336065292357,
|
| 236 |
+
"count": 5
|
| 237 |
+
},
|
| 238 |
+
"shared_aligned": {
|
| 239 |
+
"mean_shift": 0.18706549459255475,
|
| 240 |
+
"median_shift": -1.6757953744672704e-06,
|
| 241 |
+
"p95_abs_shift": 4.258874750137304,
|
| 242 |
+
"count": 1474
|
| 243 |
+
},
|
| 244 |
+
"shared_attenuated": {
|
| 245 |
+
"mean_shift": -7.682575175577572e-07,
|
| 246 |
+
"median_shift": 0.0,
|
| 247 |
+
"p95_abs_shift": 0.0,
|
| 248 |
+
"count": 4083
|
| 249 |
+
},
|
| 250 |
+
"shared_intermediate": {
|
| 251 |
+
"mean_shift": 7.5666555200435615e-06,
|
| 252 |
+
"median_shift": 0.0,
|
| 253 |
+
"p95_abs_shift": 0.0,
|
| 254 |
+
"count": 660
|
| 255 |
+
},
|
| 256 |
+
"shared_redirected": {
|
| 257 |
+
"mean_shift": 6.903093206683238e-07,
|
| 258 |
+
"median_shift": 0.0,
|
| 259 |
+
"p95_abs_shift": 0.0,
|
| 260 |
+
"count": 18354
|
| 261 |
+
}
|
| 262 |
+
},
|
| 263 |
+
"25": {
|
| 264 |
+
"aligned_only": {
|
| 265 |
+
"mean_shift": 0.11894665029558382,
|
| 266 |
+
"median_shift": 8.357907972822431e-06,
|
| 267 |
+
"p95_abs_shift": 2.852829647064209,
|
| 268 |
+
"count": 5
|
| 269 |
+
},
|
| 270 |
+
"shared_aligned": {
|
| 271 |
+
"mean_shift": 0.20790013630214754,
|
| 272 |
+
"median_shift": -8.05746958576492e-06,
|
| 273 |
+
"p95_abs_shift": 4.167022371292107,
|
| 274 |
+
"count": 1474
|
| 275 |
+
},
|
| 276 |
+
"shared_attenuated": {
|
| 277 |
+
"mean_shift": -1.924276880147058e-06,
|
| 278 |
+
"median_shift": 0.0,
|
| 279 |
+
"p95_abs_shift": 0.0,
|
| 280 |
+
"count": 4083
|
| 281 |
+
},
|
| 282 |
+
"shared_intermediate": {
|
| 283 |
+
"mean_shift": 4.6839660514179245e-06,
|
| 284 |
+
"median_shift": 0.0,
|
| 285 |
+
"p95_abs_shift": 0.0,
|
| 286 |
+
"count": 660
|
| 287 |
+
},
|
| 288 |
+
"shared_redirected": {
|
| 289 |
+
"mean_shift": 1.508880238708032e-06,
|
| 290 |
+
"median_shift": 0.0,
|
| 291 |
+
"p95_abs_shift": 0.0,
|
| 292 |
+
"count": 18354
|
| 293 |
+
}
|
| 294 |
+
}
|
| 295 |
+
},
|
| 296 |
+
"total_features": 24576,
|
| 297 |
+
"fve_base": 0.650527210759867,
|
| 298 |
+
"fve_aligned": 0.6851702085964343,
|
| 299 |
+
"fve_base_by_layer": [
|
| 300 |
+
0.6429182481391268,
|
| 301 |
+
0.6468658194492001,
|
| 302 |
+
0.661797509455556
|
| 303 |
+
],
|
| 304 |
+
"fve_aligned_by_layer": [
|
| 305 |
+
0.6799573065098667,
|
| 306 |
+
0.6851004518139425,
|
| 307 |
+
0.6904528172228349
|
| 308 |
+
],
|
| 309 |
+
"val_fve_base_by_layer": [
|
| 310 |
+
0.6429182481391268,
|
| 311 |
+
0.6468658194492001,
|
| 312 |
+
0.661797509455556
|
| 313 |
+
],
|
| 314 |
+
"val_fve_aligned_by_layer": [
|
| 315 |
+
0.6799573065098667,
|
| 316 |
+
0.6851004518139425,
|
| 317 |
+
0.6904528172228349
|
| 318 |
+
],
|
| 319 |
+
"dead_neuron_fraction": 0.9710712081775008,
|
| 320 |
+
"l0_sparsity_base": 136.68414228253383,
|
| 321 |
+
"l0_sparsity_aligned": 134.29049186461583,
|
| 322 |
+
"l0_base_by_layer": [
|
| 323 |
+
125.3520869819031,
|
| 324 |
+
134.57740440747227,
|
| 325 |
+
150.12292031523643
|
| 326 |
+
],
|
| 327 |
+
"l0_aligned_by_layer": [
|
| 328 |
+
126.03602962638645,
|
| 329 |
+
131.27096103327494,
|
| 330 |
+
145.56447022767074
|
| 331 |
+
],
|
| 332 |
+
"val_l0_base": 137.48267991630195,
|
| 333 |
+
"val_l0_aligned": 134.74987800957643,
|
| 334 |
+
"val_l0_base_by_layer": [
|
| 335 |
+
126.13634383116717,
|
| 336 |
+
135.3324607329843,
|
| 337 |
+
150.9792212041885
|
| 338 |
+
],
|
| 339 |
+
"val_l0_aligned_by_layer": [
|
| 340 |
+
126.54193938090539,
|
| 341 |
+
131.6101112565445,
|
| 342 |
+
146.0975676531567
|
| 343 |
+
],
|
| 344 |
+
"superposition_fraction": 0.0
|
| 345 |
+
}
|
v1/assembled_activations/llama32-3b-kto/metrics/training_metrics.json
ADDED
|
@@ -0,0 +1,524 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epochs": [
|
| 3 |
+
1,
|
| 4 |
+
2,
|
| 5 |
+
3,
|
| 6 |
+
4,
|
| 7 |
+
5,
|
| 8 |
+
6,
|
| 9 |
+
7,
|
| 10 |
+
8
|
| 11 |
+
],
|
| 12 |
+
"train_loss": [
|
| 13 |
+
0.5933123942327583,
|
| 14 |
+
0.442794848935895,
|
| 15 |
+
0.42607008457462203,
|
| 16 |
+
0.38806798507856893,
|
| 17 |
+
0.3780940956040843,
|
| 18 |
+
0.37194959002493144,
|
| 19 |
+
0.36910743747826424,
|
| 20 |
+
0.36662243225737334
|
| 21 |
+
],
|
| 22 |
+
"val_loss": [
|
| 23 |
+
0.4522804942430626,
|
| 24 |
+
0.43734881515902374,
|
| 25 |
+
0.41274881846617656,
|
| 26 |
+
0.38737851022425746,
|
| 27 |
+
0.38028734547929616,
|
| 28 |
+
0.3786773179214038,
|
| 29 |
+
0.3776190936253333,
|
| 30 |
+
0.37696840079667054
|
| 31 |
+
],
|
| 32 |
+
"train_fve_base": [
|
| 33 |
+
0.4814383069961313,
|
| 34 |
+
0.6003544950304293,
|
| 35 |
+
0.6189264595265007,
|
| 36 |
+
0.6378410120352763,
|
| 37 |
+
0.6485570505287518,
|
| 38 |
+
0.6568282935540021,
|
| 39 |
+
0.6628907306529733,
|
| 40 |
+
0.6666664291875654
|
| 41 |
+
],
|
| 42 |
+
"train_fve_aligned": [
|
| 43 |
+
0.4897083103253359,
|
| 44 |
+
0.5930533013842082,
|
| 45 |
+
0.6123184734662931,
|
| 46 |
+
0.6680812882667247,
|
| 47 |
+
0.680344003651202,
|
| 48 |
+
0.6861663695252993,
|
| 49 |
+
0.6893872699748683,
|
| 50 |
+
0.6921899963803133
|
| 51 |
+
],
|
| 52 |
+
"val_fve_base": [
|
| 53 |
+
0.5845968798193008,
|
| 54 |
+
0.6098661441453465,
|
| 55 |
+
0.6222324986108311,
|
| 56 |
+
0.635092594860736,
|
| 57 |
+
0.6419743776945543,
|
| 58 |
+
0.6467851612580384,
|
| 59 |
+
0.6494424973482861,
|
| 60 |
+
0.650527210759867
|
| 61 |
+
],
|
| 62 |
+
"val_fve_aligned": [
|
| 63 |
+
0.5853793015654799,
|
| 64 |
+
0.5955420280002175,
|
| 65 |
+
0.6358089334677651,
|
| 66 |
+
0.671886538023724,
|
| 67 |
+
0.680234924036795,
|
| 68 |
+
0.6819715406258069,
|
| 69 |
+
0.6842551374934731,
|
| 70 |
+
0.6851702085964343
|
| 71 |
+
],
|
| 72 |
+
"val_fve_base_by_layer": [
|
| 73 |
+
[
|
| 74 |
+
0.5732664069580158,
|
| 75 |
+
0.5763397326020046,
|
| 76 |
+
0.6041844409173696
|
| 77 |
+
],
|
| 78 |
+
[
|
| 79 |
+
0.5986687528525347,
|
| 80 |
+
0.6034028695515937,
|
| 81 |
+
0.6275267547961929
|
| 82 |
+
],
|
| 83 |
+
[
|
| 84 |
+
0.6165900526870608,
|
| 85 |
+
0.6112354472669631,
|
| 86 |
+
0.6388719418910161
|
| 87 |
+
],
|
| 88 |
+
[
|
| 89 |
+
0.6276468862413735,
|
| 90 |
+
0.6292668268942708,
|
| 91 |
+
0.6483640190194414
|
| 92 |
+
],
|
| 93 |
+
[
|
| 94 |
+
0.6339579698927116,
|
| 95 |
+
0.6365792960396612,
|
| 96 |
+
0.6553858159724331
|
| 97 |
+
],
|
| 98 |
+
[
|
| 99 |
+
0.6386778891398645,
|
| 100 |
+
0.6424399002684349,
|
| 101 |
+
0.6592376335129064
|
| 102 |
+
],
|
| 103 |
+
[
|
| 104 |
+
0.6416899889551532,
|
| 105 |
+
0.6455407226897035,
|
| 106 |
+
0.6610967317176739
|
| 107 |
+
],
|
| 108 |
+
[
|
| 109 |
+
0.6429182481391268,
|
| 110 |
+
0.6468658194492001,
|
| 111 |
+
0.661797509455556
|
| 112 |
+
]
|
| 113 |
+
],
|
| 114 |
+
"val_fve_aligned_by_layer": [
|
| 115 |
+
[
|
| 116 |
+
0.5565477425515339,
|
| 117 |
+
0.5519646746325867,
|
| 118 |
+
0.647625432276601
|
| 119 |
+
],
|
| 120 |
+
[
|
| 121 |
+
0.5665813191398901,
|
| 122 |
+
0.5605105738989345,
|
| 123 |
+
0.6595341338537126
|
| 124 |
+
],
|
| 125 |
+
[
|
| 126 |
+
0.6707772475262587,
|
| 127 |
+
0.5713420142053933,
|
| 128 |
+
0.6653074703291448
|
| 129 |
+
],
|
| 130 |
+
[
|
| 131 |
+
0.6783266744688543,
|
| 132 |
+
0.6637531731141175,
|
| 133 |
+
0.6735797000180989
|
| 134 |
+
],
|
| 135 |
+
[
|
| 136 |
+
0.6752785547241491,
|
| 137 |
+
0.6786753744355047,
|
| 138 |
+
0.6867507864667483
|
| 139 |
+
],
|
| 140 |
+
[
|
| 141 |
+
0.6770623098493247,
|
| 142 |
+
0.68111452007793,
|
| 143 |
+
0.687737735154117
|
| 144 |
+
],
|
| 145 |
+
[
|
| 146 |
+
0.6787484924830691,
|
| 147 |
+
0.6842160321655074,
|
| 148 |
+
0.689800825106536
|
| 149 |
+
],
|
| 150 |
+
[
|
| 151 |
+
0.6799573065098667,
|
| 152 |
+
0.6851004518139425,
|
| 153 |
+
0.6904528172228349
|
| 154 |
+
]
|
| 155 |
+
],
|
| 156 |
+
"train_fve_base_by_layer": [
|
| 157 |
+
[
|
| 158 |
+
0.4730913452216915,
|
| 159 |
+
0.47967155965809505,
|
| 160 |
+
0.4915519679629587
|
| 161 |
+
],
|
| 162 |
+
[
|
| 163 |
+
0.5891723040593554,
|
| 164 |
+
0.5918736373271708,
|
| 165 |
+
0.6200174819080021
|
| 166 |
+
],
|
| 167 |
+
[
|
| 168 |
+
0.6081104554840506,
|
| 169 |
+
0.6112697695474909,
|
| 170 |
+
0.6373990934909747
|
| 171 |
+
],
|
| 172 |
+
[
|
| 173 |
+
0.631499816713038,
|
| 174 |
+
0.6302260103381592,
|
| 175 |
+
0.6517971492760096
|
| 176 |
+
],
|
| 177 |
+
[
|
| 178 |
+
0.6413867825732479,
|
| 179 |
+
0.6431660363010266,
|
| 180 |
+
0.6611182747427232
|
| 181 |
+
],
|
| 182 |
+
[
|
| 183 |
+
0.6489884858234404,
|
| 184 |
+
0.651991655585784,
|
| 185 |
+
0.6695046753682939
|
| 186 |
+
],
|
| 187 |
+
[
|
| 188 |
+
0.6551149910956325,
|
| 189 |
+
0.658601807177032,
|
| 190 |
+
0.6749553329333602
|
| 191 |
+
],
|
| 192 |
+
[
|
| 193 |
+
0.6590984342951421,
|
| 194 |
+
0.6626518236846166,
|
| 195 |
+
0.6782489653504946
|
| 196 |
+
]
|
| 197 |
+
],
|
| 198 |
+
"train_fve_aligned_by_layer": [
|
| 199 |
+
[
|
| 200 |
+
0.48375090509566104,
|
| 201 |
+
0.4794864978138912,
|
| 202 |
+
0.5058874788857731
|
| 203 |
+
],
|
| 204 |
+
[
|
| 205 |
+
0.5638634942110978,
|
| 206 |
+
0.5601495567359078,
|
| 207 |
+
0.6551467945752475
|
| 208 |
+
],
|
| 209 |
+
[
|
| 210 |
+
0.6013194078669796,
|
| 211 |
+
0.5714226390490111,
|
| 212 |
+
0.6642133136694708
|
| 213 |
+
],
|
| 214 |
+
[
|
| 215 |
+
0.6761792779177583,
|
| 216 |
+
0.6555343266474317,
|
| 217 |
+
0.6725302028224542
|
| 218 |
+
],
|
| 219 |
+
[
|
| 220 |
+
0.6819632140015971,
|
| 221 |
+
0.6775305208516831,
|
| 222 |
+
0.6815382182702501
|
| 223 |
+
],
|
| 224 |
+
[
|
| 225 |
+
0.6810176338080558,
|
| 226 |
+
0.6852003142141976,
|
| 227 |
+
0.6922811020276592
|
| 228 |
+
],
|
| 229 |
+
[
|
| 230 |
+
0.6838987355544006,
|
| 231 |
+
0.6887385682341514,
|
| 232 |
+
0.6955244492454551
|
| 233 |
+
],
|
| 234 |
+
[
|
| 235 |
+
0.6868440483093818,
|
| 236 |
+
0.6917720467321927,
|
| 237 |
+
0.6979538341467657
|
| 238 |
+
]
|
| 239 |
+
],
|
| 240 |
+
"dead_neurons": [
|
| 241 |
+
0.9377401455827993,
|
| 242 |
+
0.9667631407891946,
|
| 243 |
+
0.9680073846172883,
|
| 244 |
+
0.9687811607376429,
|
| 245 |
+
0.9695221584092367,
|
| 246 |
+
0.9703716389004139,
|
| 247 |
+
0.9709274034296916,
|
| 248 |
+
0.9710712081775008
|
| 249 |
+
],
|
| 250 |
+
"l0_base": [
|
| 251 |
+
168.06772301757954,
|
| 252 |
+
126.81064258760273,
|
| 253 |
+
123.2507261019037,
|
| 254 |
+
126.34850245808136,
|
| 255 |
+
130.52682192257393,
|
| 256 |
+
132.99421610834997,
|
| 257 |
+
135.4419259128693,
|
| 258 |
+
136.68414228253383
|
| 259 |
+
],
|
| 260 |
+
"l0_aligned": [
|
| 261 |
+
159.23378112311957,
|
| 262 |
+
122.26635424630098,
|
| 263 |
+
119.10803059201038,
|
| 264 |
+
123.13043885862903,
|
| 265 |
+
126.86929286994645,
|
| 266 |
+
129.55245873695637,
|
| 267 |
+
132.93504854593093,
|
| 268 |
+
134.29049186461583
|
| 269 |
+
],
|
| 270 |
+
"l0_base_by_layer": [
|
| 271 |
+
[
|
| 272 |
+
161.7963915645067,
|
| 273 |
+
167.02606903093988,
|
| 274 |
+
175.3806917688266
|
| 275 |
+
],
|
| 276 |
+
[
|
| 277 |
+
117.48657326328079,
|
| 278 |
+
123.98622664915354,
|
| 279 |
+
138.95911777583188
|
| 280 |
+
],
|
| 281 |
+
[
|
| 282 |
+
115.22205195563339,
|
| 283 |
+
121.02900612959719,
|
| 284 |
+
133.501112813777
|
| 285 |
+
],
|
| 286 |
+
[
|
| 287 |
+
115.18999927028605,
|
| 288 |
+
123.78610259778166,
|
| 289 |
+
140.06939579684763
|
| 290 |
+
],
|
| 291 |
+
[
|
| 292 |
+
117.65137915936953,
|
| 293 |
+
129.656158785756,
|
| 294 |
+
144.27291301809692
|
| 295 |
+
],
|
| 296 |
+
[
|
| 297 |
+
120.36387186223,
|
| 298 |
+
131.73347197898423,
|
| 299 |
+
146.88528896672506
|
| 300 |
+
],
|
| 301 |
+
[
|
| 302 |
+
123.50923088149446,
|
| 303 |
+
133.5356830122592,
|
| 304 |
+
149.28084865732632
|
| 305 |
+
],
|
| 306 |
+
[
|
| 307 |
+
125.3520869819031,
|
| 308 |
+
134.57740440747227,
|
| 309 |
+
150.12292031523643
|
| 310 |
+
]
|
| 311 |
+
],
|
| 312 |
+
"l0_aligned_by_layer": [
|
| 313 |
+
[
|
| 314 |
+
155.430823117338,
|
| 315 |
+
157.74321366024517,
|
| 316 |
+
164.5272913018097
|
| 317 |
+
],
|
| 318 |
+
[
|
| 319 |
+
115.72462419731465,
|
| 320 |
+
117.97945855224752,
|
| 321 |
+
133.09497227086982
|
| 322 |
+
],
|
| 323 |
+
[
|
| 324 |
+
113.77422650321074,
|
| 325 |
+
113.63875510799767,
|
| 326 |
+
129.91110259778168
|
| 327 |
+
],
|
| 328 |
+
[
|
| 329 |
+
114.65721687098657,
|
| 330 |
+
118.46561223000583,
|
| 331 |
+
136.26848000583772
|
| 332 |
+
],
|
| 333 |
+
[
|
| 334 |
+
117.2795716579101,
|
| 335 |
+
124.29976649153532,
|
| 336 |
+
139.0285318155283
|
| 337 |
+
],
|
| 338 |
+
[
|
| 339 |
+
120.08021380618797,
|
| 340 |
+
126.71048598949211,
|
| 341 |
+
141.86666301809692
|
| 342 |
+
],
|
| 343 |
+
[
|
| 344 |
+
124.03427831290135,
|
| 345 |
+
129.95809617629888,
|
| 346 |
+
144.81275539988326
|
| 347 |
+
],
|
| 348 |
+
[
|
| 349 |
+
126.03602962638645,
|
| 350 |
+
131.27096103327494,
|
| 351 |
+
145.56447022767074
|
| 352 |
+
]
|
| 353 |
+
],
|
| 354 |
+
"val_l0_base": [
|
| 355 |
+
133.6790515040852,
|
| 356 |
+
126.03499806988302,
|
| 357 |
+
122.7813797396515,
|
| 358 |
+
129.9280155042079,
|
| 359 |
+
132.89138482378416,
|
| 360 |
+
134.8359382988895,
|
| 361 |
+
136.81023291143447,
|
| 362 |
+
137.48267991630195
|
| 363 |
+
],
|
| 364 |
+
"val_l0_aligned": [
|
| 365 |
+
127.6454003219205,
|
| 366 |
+
121.64332736349854,
|
| 367 |
+
117.30543074682745,
|
| 368 |
+
127.08719019864866,
|
| 369 |
+
128.0597404360147,
|
| 370 |
+
131.95315767457973,
|
| 371 |
+
134.10319015742596,
|
| 372 |
+
134.74987800957643
|
| 373 |
+
],
|
| 374 |
+
"val_l0_base_by_layer": [
|
| 375 |
+
[
|
| 376 |
+
125.309500449615,
|
| 377 |
+
128.9821662303665,
|
| 378 |
+
146.745473412319
|
| 379 |
+
],
|
| 380 |
+
[
|
| 381 |
+
115.72027707724047,
|
| 382 |
+
125.55879147514622,
|
| 383 |
+
136.8259162303665
|
| 384 |
+
],
|
| 385 |
+
[
|
| 386 |
+
113.56468152750224,
|
| 387 |
+
120.74323736060977,
|
| 388 |
+
134.03621294111483
|
| 389 |
+
],
|
| 390 |
+
[
|
| 391 |
+
117.10433029254693,
|
| 392 |
+
127.62696335078535,
|
| 393 |
+
145.05273781022476
|
| 394 |
+
],
|
| 395 |
+
[
|
| 396 |
+
120.67233859556508,
|
| 397 |
+
130.8184446165075,
|
| 398 |
+
147.18335520160136
|
| 399 |
+
],
|
| 400 |
+
[
|
| 401 |
+
122.73429319371728,
|
| 402 |
+
132.8148451400677,
|
| 403 |
+
148.95866058509387
|
| 404 |
+
],
|
| 405 |
+
[
|
| 406 |
+
125.38885254385583,
|
| 407 |
+
134.5613547120419,
|
| 408 |
+
150.48047562044954
|
| 409 |
+
],
|
| 410 |
+
[
|
| 411 |
+
126.13634383116717,
|
| 412 |
+
135.3324607329843,
|
| 413 |
+
150.9792212041885
|
| 414 |
+
]
|
| 415 |
+
],
|
| 416 |
+
"val_l0_aligned_by_layer": [
|
| 417 |
+
[
|
| 418 |
+
124.38279887893437,
|
| 419 |
+
120.35596641820139,
|
| 420 |
+
138.19742588222962
|
| 421 |
+
],
|
| 422 |
+
[
|
| 423 |
+
114.432973412319,
|
| 424 |
+
117.49503709882966,
|
| 425 |
+
133.00196335078533
|
| 426 |
+
],
|
| 427 |
+
[
|
| 428 |
+
109.7994109947644,
|
| 429 |
+
110.60089442867259,
|
| 430 |
+
131.5159795471511
|
| 431 |
+
],
|
| 432 |
+
[
|
| 433 |
+
116.04553885734518,
|
| 434 |
+
124.60585732984293,
|
| 435 |
+
140.61016582069595
|
| 436 |
+
],
|
| 437 |
+
[
|
| 438 |
+
119.92277486910994,
|
| 439 |
+
124.03735822907294,
|
| 440 |
+
140.2190772251309
|
| 441 |
+
],
|
| 442 |
+
[
|
| 443 |
+
122.63519854320906,
|
| 444 |
+
129.07717060668307,
|
| 445 |
+
144.1470876963351
|
| 446 |
+
],
|
| 447 |
+
[
|
| 448 |
+
125.78043193717278,
|
| 449 |
+
130.91977535866943,
|
| 450 |
+
145.60934775786873
|
| 451 |
+
],
|
| 452 |
+
[
|
| 453 |
+
126.54193938090539,
|
| 454 |
+
131.6101112565445,
|
| 455 |
+
146.0975676531567
|
| 456 |
+
]
|
| 457 |
+
],
|
| 458 |
+
"self_recon": [
|
| 459 |
+
0.3596775165024516,
|
| 460 |
+
0.2799183193418065,
|
| 461 |
+
0.26739436908208064,
|
| 462 |
+
0.24186065978323307,
|
| 463 |
+
0.23369103840618055,
|
| 464 |
+
0.22850491952910038,
|
| 465 |
+
0.22538104173180937,
|
| 466 |
+
0.22313513932559342
|
| 467 |
+
],
|
| 468 |
+
"cross_recon": [
|
| 469 |
+
0.38041265005266467,
|
| 470 |
+
0.3039850652113479,
|
| 471 |
+
0.29148193161284375,
|
| 472 |
+
0.26661263893413323,
|
| 473 |
+
0.2592871804127412,
|
| 474 |
+
0.25357689010387124,
|
| 475 |
+
0.2501190767187997,
|
| 476 |
+
0.24781497845820913
|
| 477 |
+
],
|
| 478 |
+
"sparsity": [
|
| 479 |
+
0.0814698132715132,
|
| 480 |
+
0.04128250095208207,
|
| 481 |
+
0.04208294076924427,
|
| 482 |
+
0.03956226723554603,
|
| 483 |
+
0.04068818295856212,
|
| 484 |
+
0.04201391230131267,
|
| 485 |
+
0.04367876260541923,
|
| 486 |
+
0.04436129915717744
|
| 487 |
+
],
|
| 488 |
+
"val_self_recon": [
|
| 489 |
+
0.28703898846791054,
|
| 490 |
+
0.27524666722220276,
|
| 491 |
+
0.25847701931186995,
|
| 492 |
+
0.2406287566997618,
|
| 493 |
+
0.23478099087458007,
|
| 494 |
+
0.2326575024589818,
|
| 495 |
+
0.23096176958521,
|
| 496 |
+
0.23029655413165767
|
| 497 |
+
],
|
| 498 |
+
"val_cross_recon": [
|
| 499 |
+
0.311376446516726,
|
| 500 |
+
0.29898168794147634,
|
| 501 |
+
0.2829330632979957,
|
| 502 |
+
0.26673236247444654,
|
| 503 |
+
0.2611719256764307,
|
| 504 |
+
0.25862815227183994,
|
| 505 |
+
0.2567455348856162,
|
| 506 |
+
0.2561143940187874
|
| 507 |
+
],
|
| 508 |
+
"val_sparsity": [
|
| 509 |
+
0.04069092366551853,
|
| 510 |
+
0.042509469140262504,
|
| 511 |
+
0.04109857126056212,
|
| 512 |
+
0.04005680582364192,
|
| 513 |
+
0.04103758367490394,
|
| 514 |
+
0.04256855082059406,
|
| 515 |
+
0.04395910826419037,
|
| 516 |
+
0.044226086713100604
|
| 517 |
+
],
|
| 518 |
+
"layers": [
|
| 519 |
+
23,
|
| 520 |
+
24,
|
| 521 |
+
25
|
| 522 |
+
],
|
| 523 |
+
"topk_mode": "model_balanced_layer_agg"
|
| 524 |
+
}
|
v1/assembled_activations/llama32-3b-kto/plots/aligned_decoder_norm_heatmap.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/base_decoder_norm_heatmap.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/cf_shift_by_layer.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/cf_shift_p95_by_layer.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/class_distribution_multilayer.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/class_distribution_primary.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/cross_layer_cosine_drift_by_stream.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/decoder_norm_ratio_by_layer.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/feature_layer_trajectories.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/feature_sharing_ratio_by_layer.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/fve_by_layer.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/l0_by_layer.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/layer_concentration_entropy.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/loss_curves.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/max_norm_layer_migration.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/rho_histogram_by_layer.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/rho_theta_scatter_by_layer.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/superposition_by_layer.png
ADDED
|
Git LFS Details
|
v1/assembled_activations/llama32-3b-kto/plots/theta_by_layer.png
ADDED
|
Git LFS Details
|