nvan15 commited on
Commit
e6dd826
·
verified ·
1 Parent(s): b3ccc92

Batch upload part 16

Browse files
Files changed (50) hide show
  1. nl_tasks/exp100/run_ex01/all_results.json +17 -0
  2. nl_tasks/exp100/run_ex01/eval_results.json +9 -0
  3. nl_tasks/exp100/run_ex01/ft/adapter_config.json +22 -0
  4. nl_tasks/exp100/run_ex01/ft/added_tokens.json +3 -0
  5. nl_tasks/exp100/run_ex01/ft/special_tokens_map.json +15 -0
  6. nl_tasks/exp100/run_ex01/ft/spm.model +3 -0
  7. nl_tasks/exp100/run_ex01/ft/tokenizer.json +0 -0
  8. nl_tasks/exp100/run_ex01/ft/tokenizer.model +3 -0
  9. nl_tasks/exp100/run_ex01/ft/tokenizer_config.json +60 -0
  10. nl_tasks/exp100/run_ex01/ft2/adapter_config.json +22 -0
  11. nl_tasks/exp100/run_ex01/ft2/adapter_model.bin +3 -0
  12. nl_tasks/exp100/run_ex01/predict_results_cola.txt +301 -0
  13. nl_tasks/exp100/run_ex01/predict_results_mnli-mm.txt +0 -0
  14. nl_tasks/exp100/run_ex01/predict_results_mnli.txt +301 -0
  15. nl_tasks/exp100/run_ex01/trainer_state.json +43 -0
  16. nl_tasks/exp100/run_ex02/trainer_state.json +260 -0
  17. nl_tasks/exp395/run_ex09/ft/adapter_config.json +18 -0
  18. nl_tasks/exp395/run_ex09/ft/special_tokens_map.json +24 -0
  19. nl_tasks/exp395/run_ex09/ft/tokenizer.json +0 -0
  20. nl_tasks/exp395/run_ex09/ft/tokenizer.model +3 -0
  21. nl_tasks/exp395/run_ex09/ft/tokenizer_config.json +43 -0
  22. nl_tasks/exp395/run_ex09/ft2/adapter_config.json +18 -0
  23. nl_tasks/exp395/run_ex09/ft2/adapter_model.bin +3 -0
  24. nl_tasks/exp395/run_ex09/trainer_state.json +509 -0
  25. nl_tasks/exp395/run_ex10/ft/adapter_config.json +18 -0
  26. nl_tasks/exp395/run_ex10/ft/special_tokens_map.json +24 -0
  27. nl_tasks/exp395/run_ex10/ft/tokenizer.json +0 -0
  28. nl_tasks/exp395/run_ex10/ft/tokenizer.model +3 -0
  29. nl_tasks/exp395/run_ex10/ft/tokenizer_config.json +43 -0
  30. nl_tasks/exp395/run_ex10/ft2/adapter_config.json +18 -0
  31. nl_tasks/exp395/run_ex10/ft2/adapter_model.bin +3 -0
  32. nl_tasks/exp395/run_ex10/trainer_state.json +356 -0
  33. nl_tasks/exp_init/run_ex01/ft/adapter_config.json +18 -0
  34. nl_tasks/exp_init/run_ex01/ft/added_tokens.json +3 -0
  35. nl_tasks/exp_init/run_ex01/ft/special_tokens_map.json +30 -0
  36. nl_tasks/exp_init/run_ex01/ft/tokenizer.json +0 -0
  37. nl_tasks/exp_init/run_ex01/ft/tokenizer.model +3 -0
  38. nl_tasks/exp_init/run_ex01/ft/tokenizer_config.json +51 -0
  39. nl_tasks/exp_init/run_ex01/ft2/adapter_config.json +18 -0
  40. nl_tasks/exp_init/run_ex01/ft2/adapter_model.bin +3 -0
  41. nl_tasks/exp_init/run_ex01/trainer_state.json +928 -0
  42. nl_tasks/exp_init/run_ex02/ft/adapter_config.json +18 -0
  43. nl_tasks/exp_init/run_ex02/ft/added_tokens.json +3 -0
  44. nl_tasks/exp_init/run_ex02/ft/special_tokens_map.json +30 -0
  45. nl_tasks/exp_init/run_ex02/ft/tokenizer.json +0 -0
  46. nl_tasks/exp_init/run_ex02/ft/tokenizer.model +3 -0
  47. nl_tasks/exp_init/run_ex02/ft/tokenizer_config.json +51 -0
  48. nl_tasks/exp_init/run_ex02/ft2/adapter_config.json +18 -0
  49. nl_tasks/exp_init/run_ex02/ft2/adapter_model.bin +3 -0
  50. nl_tasks/exp_init/run_ex02/trainer_state.json +914 -0
nl_tasks/exp100/run_ex01/all_results.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "epoch_mm": 2.0,
4
+ "eval_accuracy": 0.3273560876209883,
5
+ "eval_accuracy_mm": 0.3273560876209883,
6
+ "eval_loss": 0.5474581718444824,
7
+ "eval_loss_mm": 1.1032307147979736,
8
+ "eval_matthews_correlation": 0.10385466804655791,
9
+ "eval_runtime": 5.1907,
10
+ "eval_runtime_mm": 45.7755,
11
+ "eval_samples": 1043,
12
+ "eval_samples_mm": 9832,
13
+ "eval_samples_per_second": 200.936,
14
+ "eval_samples_per_second_mm": 214.416,
15
+ "eval_steps_per_second": 3.275,
16
+ "eval_steps_per_second_mm": 3.364
17
+ }
nl_tasks/exp100/run_ex01/eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "eval_loss": 0.5474581718444824,
4
+ "eval_matthews_correlation": 0.10385466804655791,
5
+ "eval_runtime": 5.1907,
6
+ "eval_samples": 1043,
7
+ "eval_samples_per_second": 200.936,
8
+ "eval_steps_per_second": 3.275
9
+ }
nl_tasks/exp100/run_ex01/ft/adapter_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "T": 1.0,
3
+ "base_model_name_or_path": "microsoft/deberta-v3-base",
4
+ "bias": "none",
5
+ "inference_mode": false,
6
+ "layers_to_transform": null,
7
+ "modules_to_save": null,
8
+ "num_rotations": 1,
9
+ "peft_type": "ROTATION",
10
+ "r": 5,
11
+ "revision": null,
12
+ "target_modules": [
13
+ "query_proj",
14
+ "attention.output.dense",
15
+ "output.dense",
16
+ "value_proj",
17
+ "intermediate.dense",
18
+ "key_proj"
19
+ ],
20
+ "target_modules_to_skip": null,
21
+ "task_type": "SEQ_CLS"
22
+ }
nl_tasks/exp100/run_ex01/ft/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
nl_tasks/exp100/run_ex01/ft/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
nl_tasks/exp100/run_ex01/ft/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
nl_tasks/exp100/run_ex01/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
nl_tasks/exp100/run_ex01/ft/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
nl_tasks/exp100/run_ex01/ft/tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": false,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "extra_special_tokens": {},
50
+ "mask_token": "[MASK]",
51
+ "model_max_length": 512,
52
+ "pad_token": "[PAD]",
53
+ "padding_side": "right",
54
+ "sep_token": "[SEP]",
55
+ "sp_model_kwargs": {},
56
+ "split_by_punct": false,
57
+ "tokenizer_class": "DebertaV2Tokenizer",
58
+ "unk_token": "[UNK]",
59
+ "vocab_type": "spm"
60
+ }
nl_tasks/exp100/run_ex01/ft2/adapter_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "T": 1.0,
3
+ "base_model_name_or_path": "microsoft/deberta-v3-base",
4
+ "bias": "none",
5
+ "inference_mode": true,
6
+ "layers_to_transform": null,
7
+ "modules_to_save": null,
8
+ "num_rotations": 1,
9
+ "peft_type": "ROTATION",
10
+ "r": 5,
11
+ "revision": null,
12
+ "target_modules": [
13
+ "query_proj",
14
+ "attention.output.dense",
15
+ "output.dense",
16
+ "value_proj",
17
+ "intermediate.dense",
18
+ "key_proj"
19
+ ],
20
+ "target_modules_to_skip": null,
21
+ "task_type": "SEQ_CLS"
22
+ }
nl_tasks/exp100/run_ex01/ft2/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de19a85f9671833d9995e7405021a6fa84f03b7b2108e4224f853c11f3caa47c
3
+ size 3380395
nl_tasks/exp100/run_ex01/predict_results_cola.txt ADDED
@@ -0,0 +1,301 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ index prediction
2
+ 0 acceptable
3
+ 1 acceptable
4
+ 2 acceptable
5
+ 3 acceptable
6
+ 4 acceptable
7
+ 5 acceptable
8
+ 6 acceptable
9
+ 7 acceptable
10
+ 8 acceptable
11
+ 9 acceptable
12
+ 10 acceptable
13
+ 11 acceptable
14
+ 12 acceptable
15
+ 13 acceptable
16
+ 14 acceptable
17
+ 15 acceptable
18
+ 16 acceptable
19
+ 17 acceptable
20
+ 18 unacceptable
21
+ 19 acceptable
22
+ 20 acceptable
23
+ 21 acceptable
24
+ 22 acceptable
25
+ 23 acceptable
26
+ 24 acceptable
27
+ 25 acceptable
28
+ 26 acceptable
29
+ 27 acceptable
30
+ 28 acceptable
31
+ 29 acceptable
32
+ 30 acceptable
33
+ 31 acceptable
34
+ 32 acceptable
35
+ 33 acceptable
36
+ 34 acceptable
37
+ 35 acceptable
38
+ 36 acceptable
39
+ 37 acceptable
40
+ 38 acceptable
41
+ 39 acceptable
42
+ 40 acceptable
43
+ 41 acceptable
44
+ 42 acceptable
45
+ 43 acceptable
46
+ 44 acceptable
47
+ 45 acceptable
48
+ 46 acceptable
49
+ 47 acceptable
50
+ 48 acceptable
51
+ 49 acceptable
52
+ 50 acceptable
53
+ 51 acceptable
54
+ 52 acceptable
55
+ 53 acceptable
56
+ 54 acceptable
57
+ 55 acceptable
58
+ 56 acceptable
59
+ 57 acceptable
60
+ 58 acceptable
61
+ 59 acceptable
62
+ 60 acceptable
63
+ 61 acceptable
64
+ 62 acceptable
65
+ 63 acceptable
66
+ 64 acceptable
67
+ 65 acceptable
68
+ 66 acceptable
69
+ 67 acceptable
70
+ 68 acceptable
71
+ 69 acceptable
72
+ 70 acceptable
73
+ 71 acceptable
74
+ 72 acceptable
75
+ 73 acceptable
76
+ 74 acceptable
77
+ 75 acceptable
78
+ 76 acceptable
79
+ 77 acceptable
80
+ 78 acceptable
81
+ 79 acceptable
82
+ 80 acceptable
83
+ 81 acceptable
84
+ 82 acceptable
85
+ 83 acceptable
86
+ 84 acceptable
87
+ 85 acceptable
88
+ 86 acceptable
89
+ 87 acceptable
90
+ 88 acceptable
91
+ 89 acceptable
92
+ 90 acceptable
93
+ 91 acceptable
94
+ 92 acceptable
95
+ 93 acceptable
96
+ 94 acceptable
97
+ 95 acceptable
98
+ 96 acceptable
99
+ 97 acceptable
100
+ 98 acceptable
101
+ 99 acceptable
102
+ 100 acceptable
103
+ 101 acceptable
104
+ 102 acceptable
105
+ 103 acceptable
106
+ 104 acceptable
107
+ 105 acceptable
108
+ 106 acceptable
109
+ 107 acceptable
110
+ 108 acceptable
111
+ 109 acceptable
112
+ 110 acceptable
113
+ 111 acceptable
114
+ 112 acceptable
115
+ 113 acceptable
116
+ 114 acceptable
117
+ 115 acceptable
118
+ 116 acceptable
119
+ 117 acceptable
120
+ 118 acceptable
121
+ 119 acceptable
122
+ 120 acceptable
123
+ 121 acceptable
124
+ 122 acceptable
125
+ 123 acceptable
126
+ 124 acceptable
127
+ 125 acceptable
128
+ 126 acceptable
129
+ 127 acceptable
130
+ 128 acceptable
131
+ 129 acceptable
132
+ 130 acceptable
133
+ 131 acceptable
134
+ 132 acceptable
135
+ 133 acceptable
136
+ 134 acceptable
137
+ 135 acceptable
138
+ 136 acceptable
139
+ 137 acceptable
140
+ 138 acceptable
141
+ 139 acceptable
142
+ 140 acceptable
143
+ 141 acceptable
144
+ 142 acceptable
145
+ 143 acceptable
146
+ 144 acceptable
147
+ 145 acceptable
148
+ 146 acceptable
149
+ 147 acceptable
150
+ 148 acceptable
151
+ 149 acceptable
152
+ 150 acceptable
153
+ 151 acceptable
154
+ 152 acceptable
155
+ 153 acceptable
156
+ 154 acceptable
157
+ 155 acceptable
158
+ 156 acceptable
159
+ 157 acceptable
160
+ 158 acceptable
161
+ 159 acceptable
162
+ 160 acceptable
163
+ 161 acceptable
164
+ 162 acceptable
165
+ 163 acceptable
166
+ 164 acceptable
167
+ 165 acceptable
168
+ 166 acceptable
169
+ 167 acceptable
170
+ 168 acceptable
171
+ 169 acceptable
172
+ 170 acceptable
173
+ 171 acceptable
174
+ 172 acceptable
175
+ 173 acceptable
176
+ 174 acceptable
177
+ 175 acceptable
178
+ 176 acceptable
179
+ 177 acceptable
180
+ 178 acceptable
181
+ 179 acceptable
182
+ 180 acceptable
183
+ 181 acceptable
184
+ 182 acceptable
185
+ 183 acceptable
186
+ 184 acceptable
187
+ 185 acceptable
188
+ 186 acceptable
189
+ 187 acceptable
190
+ 188 acceptable
191
+ 189 acceptable
192
+ 190 acceptable
193
+ 191 acceptable
194
+ 192 acceptable
195
+ 193 acceptable
196
+ 194 acceptable
197
+ 195 acceptable
198
+ 196 acceptable
199
+ 197 acceptable
200
+ 198 acceptable
201
+ 199 acceptable
202
+ 200 acceptable
203
+ 201 acceptable
204
+ 202 acceptable
205
+ 203 acceptable
206
+ 204 acceptable
207
+ 205 acceptable
208
+ 206 acceptable
209
+ 207 acceptable
210
+ 208 acceptable
211
+ 209 acceptable
212
+ 210 acceptable
213
+ 211 acceptable
214
+ 212 acceptable
215
+ 213 acceptable
216
+ 214 acceptable
217
+ 215 acceptable
218
+ 216 acceptable
219
+ 217 acceptable
220
+ 218 acceptable
221
+ 219 acceptable
222
+ 220 acceptable
223
+ 221 acceptable
224
+ 222 acceptable
225
+ 223 acceptable
226
+ 224 acceptable
227
+ 225 acceptable
228
+ 226 acceptable
229
+ 227 acceptable
230
+ 228 acceptable
231
+ 229 acceptable
232
+ 230 acceptable
233
+ 231 acceptable
234
+ 232 acceptable
235
+ 233 acceptable
236
+ 234 acceptable
237
+ 235 acceptable
238
+ 236 acceptable
239
+ 237 acceptable
240
+ 238 acceptable
241
+ 239 acceptable
242
+ 240 acceptable
243
+ 241 acceptable
244
+ 242 acceptable
245
+ 243 acceptable
246
+ 244 acceptable
247
+ 245 acceptable
248
+ 246 acceptable
249
+ 247 acceptable
250
+ 248 acceptable
251
+ 249 acceptable
252
+ 250 acceptable
253
+ 251 acceptable
254
+ 252 acceptable
255
+ 253 acceptable
256
+ 254 acceptable
257
+ 255 acceptable
258
+ 256 acceptable
259
+ 257 acceptable
260
+ 258 acceptable
261
+ 259 acceptable
262
+ 260 acceptable
263
+ 261 acceptable
264
+ 262 acceptable
265
+ 263 acceptable
266
+ 264 acceptable
267
+ 265 acceptable
268
+ 266 acceptable
269
+ 267 acceptable
270
+ 268 acceptable
271
+ 269 acceptable
272
+ 270 acceptable
273
+ 271 acceptable
274
+ 272 acceptable
275
+ 273 acceptable
276
+ 274 acceptable
277
+ 275 acceptable
278
+ 276 acceptable
279
+ 277 acceptable
280
+ 278 acceptable
281
+ 279 acceptable
282
+ 280 acceptable
283
+ 281 acceptable
284
+ 282 acceptable
285
+ 283 acceptable
286
+ 284 acceptable
287
+ 285 acceptable
288
+ 286 acceptable
289
+ 287 acceptable
290
+ 288 acceptable
291
+ 289 acceptable
292
+ 290 acceptable
293
+ 291 acceptable
294
+ 292 acceptable
295
+ 293 acceptable
296
+ 294 acceptable
297
+ 295 acceptable
298
+ 296 acceptable
299
+ 297 acceptable
300
+ 298 acceptable
301
+ 299 acceptable
nl_tasks/exp100/run_ex01/predict_results_mnli-mm.txt ADDED
The diff for this file is too large to render. See raw diff
 
nl_tasks/exp100/run_ex01/predict_results_mnli.txt ADDED
@@ -0,0 +1,301 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ index prediction
2
+ 0 contradiction
3
+ 1 contradiction
4
+ 2 contradiction
5
+ 3 contradiction
6
+ 4 contradiction
7
+ 5 contradiction
8
+ 6 contradiction
9
+ 7 contradiction
10
+ 8 contradiction
11
+ 9 contradiction
12
+ 10 contradiction
13
+ 11 contradiction
14
+ 12 contradiction
15
+ 13 contradiction
16
+ 14 contradiction
17
+ 15 contradiction
18
+ 16 contradiction
19
+ 17 contradiction
20
+ 18 contradiction
21
+ 19 contradiction
22
+ 20 contradiction
23
+ 21 contradiction
24
+ 22 contradiction
25
+ 23 contradiction
26
+ 24 contradiction
27
+ 25 contradiction
28
+ 26 contradiction
29
+ 27 contradiction
30
+ 28 contradiction
31
+ 29 contradiction
32
+ 30 contradiction
33
+ 31 contradiction
34
+ 32 contradiction
35
+ 33 contradiction
36
+ 34 contradiction
37
+ 35 contradiction
38
+ 36 contradiction
39
+ 37 contradiction
40
+ 38 contradiction
41
+ 39 contradiction
42
+ 40 contradiction
43
+ 41 contradiction
44
+ 42 contradiction
45
+ 43 contradiction
46
+ 44 contradiction
47
+ 45 contradiction
48
+ 46 contradiction
49
+ 47 contradiction
50
+ 48 contradiction
51
+ 49 contradiction
52
+ 50 contradiction
53
+ 51 contradiction
54
+ 52 contradiction
55
+ 53 contradiction
56
+ 54 contradiction
57
+ 55 contradiction
58
+ 56 contradiction
59
+ 57 contradiction
60
+ 58 contradiction
61
+ 59 contradiction
62
+ 60 contradiction
63
+ 61 contradiction
64
+ 62 contradiction
65
+ 63 contradiction
66
+ 64 contradiction
67
+ 65 contradiction
68
+ 66 contradiction
69
+ 67 contradiction
70
+ 68 contradiction
71
+ 69 contradiction
72
+ 70 contradiction
73
+ 71 contradiction
74
+ 72 contradiction
75
+ 73 contradiction
76
+ 74 contradiction
77
+ 75 contradiction
78
+ 76 contradiction
79
+ 77 contradiction
80
+ 78 contradiction
81
+ 79 contradiction
82
+ 80 contradiction
83
+ 81 contradiction
84
+ 82 contradiction
85
+ 83 contradiction
86
+ 84 contradiction
87
+ 85 contradiction
88
+ 86 contradiction
89
+ 87 contradiction
90
+ 88 contradiction
91
+ 89 contradiction
92
+ 90 contradiction
93
+ 91 contradiction
94
+ 92 contradiction
95
+ 93 contradiction
96
+ 94 contradiction
97
+ 95 contradiction
98
+ 96 contradiction
99
+ 97 contradiction
100
+ 98 contradiction
101
+ 99 contradiction
102
+ 100 contradiction
103
+ 101 contradiction
104
+ 102 contradiction
105
+ 103 contradiction
106
+ 104 contradiction
107
+ 105 contradiction
108
+ 106 contradiction
109
+ 107 contradiction
110
+ 108 contradiction
111
+ 109 contradiction
112
+ 110 contradiction
113
+ 111 contradiction
114
+ 112 contradiction
115
+ 113 contradiction
116
+ 114 contradiction
117
+ 115 contradiction
118
+ 116 contradiction
119
+ 117 contradiction
120
+ 118 contradiction
121
+ 119 contradiction
122
+ 120 contradiction
123
+ 121 contradiction
124
+ 122 contradiction
125
+ 123 contradiction
126
+ 124 contradiction
127
+ 125 contradiction
128
+ 126 contradiction
129
+ 127 contradiction
130
+ 128 contradiction
131
+ 129 contradiction
132
+ 130 contradiction
133
+ 131 contradiction
134
+ 132 contradiction
135
+ 133 contradiction
136
+ 134 contradiction
137
+ 135 contradiction
138
+ 136 contradiction
139
+ 137 contradiction
140
+ 138 contradiction
141
+ 139 contradiction
142
+ 140 contradiction
143
+ 141 contradiction
144
+ 142 contradiction
145
+ 143 contradiction
146
+ 144 contradiction
147
+ 145 contradiction
148
+ 146 contradiction
149
+ 147 contradiction
150
+ 148 contradiction
151
+ 149 contradiction
152
+ 150 contradiction
153
+ 151 contradiction
154
+ 152 contradiction
155
+ 153 contradiction
156
+ 154 contradiction
157
+ 155 contradiction
158
+ 156 contradiction
159
+ 157 contradiction
160
+ 158 contradiction
161
+ 159 contradiction
162
+ 160 contradiction
163
+ 161 contradiction
164
+ 162 contradiction
165
+ 163 contradiction
166
+ 164 contradiction
167
+ 165 contradiction
168
+ 166 contradiction
169
+ 167 contradiction
170
+ 168 contradiction
171
+ 169 contradiction
172
+ 170 contradiction
173
+ 171 contradiction
174
+ 172 contradiction
175
+ 173 contradiction
176
+ 174 contradiction
177
+ 175 contradiction
178
+ 176 contradiction
179
+ 177 contradiction
180
+ 178 contradiction
181
+ 179 contradiction
182
+ 180 contradiction
183
+ 181 contradiction
184
+ 182 contradiction
185
+ 183 contradiction
186
+ 184 contradiction
187
+ 185 contradiction
188
+ 186 contradiction
189
+ 187 contradiction
190
+ 188 contradiction
191
+ 189 contradiction
192
+ 190 contradiction
193
+ 191 contradiction
194
+ 192 contradiction
195
+ 193 contradiction
196
+ 194 contradiction
197
+ 195 contradiction
198
+ 196 contradiction
199
+ 197 contradiction
200
+ 198 contradiction
201
+ 199 contradiction
202
+ 200 contradiction
203
+ 201 contradiction
204
+ 202 contradiction
205
+ 203 contradiction
206
+ 204 contradiction
207
+ 205 contradiction
208
+ 206 contradiction
209
+ 207 contradiction
210
+ 208 contradiction
211
+ 209 contradiction
212
+ 210 contradiction
213
+ 211 contradiction
214
+ 212 contradiction
215
+ 213 contradiction
216
+ 214 contradiction
217
+ 215 contradiction
218
+ 216 contradiction
219
+ 217 contradiction
220
+ 218 contradiction
221
+ 219 contradiction
222
+ 220 contradiction
223
+ 221 contradiction
224
+ 222 contradiction
225
+ 223 contradiction
226
+ 224 contradiction
227
+ 225 contradiction
228
+ 226 contradiction
229
+ 227 contradiction
230
+ 228 contradiction
231
+ 229 contradiction
232
+ 230 contradiction
233
+ 231 contradiction
234
+ 232 contradiction
235
+ 233 contradiction
236
+ 234 contradiction
237
+ 235 contradiction
238
+ 236 contradiction
239
+ 237 contradiction
240
+ 238 contradiction
241
+ 239 contradiction
242
+ 240 contradiction
243
+ 241 contradiction
244
+ 242 contradiction
245
+ 243 contradiction
246
+ 244 contradiction
247
+ 245 contradiction
248
+ 246 contradiction
249
+ 247 contradiction
250
+ 248 contradiction
251
+ 249 contradiction
252
+ 250 contradiction
253
+ 251 contradiction
254
+ 252 contradiction
255
+ 253 contradiction
256
+ 254 contradiction
257
+ 255 contradiction
258
+ 256 contradiction
259
+ 257 contradiction
260
+ 258 contradiction
261
+ 259 contradiction
262
+ 260 contradiction
263
+ 261 contradiction
264
+ 262 contradiction
265
+ 263 contradiction
266
+ 264 contradiction
267
+ 265 contradiction
268
+ 266 contradiction
269
+ 267 contradiction
270
+ 268 contradiction
271
+ 269 contradiction
272
+ 270 contradiction
273
+ 271 contradiction
274
+ 272 contradiction
275
+ 273 contradiction
276
+ 274 contradiction
277
+ 275 contradiction
278
+ 276 contradiction
279
+ 277 contradiction
280
+ 278 contradiction
281
+ 279 contradiction
282
+ 280 contradiction
283
+ 281 contradiction
284
+ 282 contradiction
285
+ 283 contradiction
286
+ 284 contradiction
287
+ 285 contradiction
288
+ 286 contradiction
289
+ 287 contradiction
290
+ 288 contradiction
291
+ 289 contradiction
292
+ 290 contradiction
293
+ 291 contradiction
294
+ 292 contradiction
295
+ 293 contradiction
296
+ 294 contradiction
297
+ 295 contradiction
298
+ 296 contradiction
299
+ 297 contradiction
300
+ 298 contradiction
301
+ 299 contradiction
nl_tasks/exp100/run_ex01/trainer_state.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 20,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 2.0,
14
+ "step": 20,
15
+ "total_flos": 79699144089600.0,
16
+ "train_loss": 0.6215126037597656,
17
+ "train_runtime": 9.6469,
18
+ "train_samples_per_second": 62.196,
19
+ "train_steps_per_second": 2.073
20
+ }
21
+ ],
22
+ "logging_steps": 200,
23
+ "max_steps": 20,
24
+ "num_input_tokens_seen": 0,
25
+ "num_train_epochs": 2,
26
+ "save_steps": 0,
27
+ "stateful_callbacks": {
28
+ "TrainerControl": {
29
+ "args": {
30
+ "should_epoch_stop": false,
31
+ "should_evaluate": false,
32
+ "should_log": false,
33
+ "should_save": true,
34
+ "should_training_stop": true
35
+ },
36
+ "attributes": {}
37
+ }
38
+ },
39
+ "total_flos": 79699144089600.0,
40
+ "train_batch_size": 32,
41
+ "trial_name": null,
42
+ "trial_params": null
43
+ }
nl_tasks/exp100/run_ex02/trainer_state.json ADDED
@@ -0,0 +1,260 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 6250,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.064,
14
+ "grad_norm": 0.0752563327550888,
15
+ "learning_rate": 0.01997616502308297,
16
+ "loss": 0.7328,
17
+ "step": 200
18
+ },
19
+ {
20
+ "epoch": 0.128,
21
+ "grad_norm": 0.04924536868929863,
22
+ "learning_rate": 0.0198548107876105,
23
+ "loss": 0.2951,
24
+ "step": 400
25
+ },
26
+ {
27
+ "epoch": 0.192,
28
+ "grad_norm": 0.03875705227255821,
29
+ "learning_rate": 0.01963190781883806,
30
+ "loss": 0.2758,
31
+ "step": 600
32
+ },
33
+ {
34
+ "epoch": 0.256,
35
+ "grad_norm": 0.04360447824001312,
36
+ "learning_rate": 0.019309753016687477,
37
+ "loss": 0.2649,
38
+ "step": 800
39
+ },
40
+ {
41
+ "epoch": 0.32,
42
+ "grad_norm": 0.04730730503797531,
43
+ "learning_rate": 0.01889166601994126,
44
+ "loss": 0.2546,
45
+ "step": 1000
46
+ },
47
+ {
48
+ "epoch": 0.384,
49
+ "grad_norm": 0.030088823288679123,
50
+ "learning_rate": 0.01838195499908018,
51
+ "loss": 0.2467,
52
+ "step": 1200
53
+ },
54
+ {
55
+ "epoch": 0.448,
56
+ "grad_norm": 0.03133947774767876,
57
+ "learning_rate": 0.017785872262812443,
58
+ "loss": 0.2418,
59
+ "step": 1400
60
+ },
61
+ {
62
+ "epoch": 0.512,
63
+ "grad_norm": 0.027547890320420265,
64
+ "learning_rate": 0.017109560135746254,
65
+ "loss": 0.2404,
66
+ "step": 1600
67
+ },
68
+ {
69
+ "epoch": 0.576,
70
+ "grad_norm": 0.028105350211262703,
71
+ "learning_rate": 0.01635998766490841,
72
+ "loss": 0.2368,
73
+ "step": 1800
74
+ },
75
+ {
76
+ "epoch": 0.64,
77
+ "grad_norm": 0.019630730152130127,
78
+ "learning_rate": 0.015544878807315494,
79
+ "loss": 0.2364,
80
+ "step": 2000
81
+ },
82
+ {
83
+ "epoch": 0.704,
84
+ "grad_norm": 0.022615477442741394,
85
+ "learning_rate": 0.014672632838587716,
86
+ "loss": 0.23,
87
+ "step": 2200
88
+ },
89
+ {
90
+ "epoch": 0.768,
91
+ "grad_norm": 0.018555844202637672,
92
+ "learning_rate": 0.01375223780275345,
93
+ "loss": 0.2271,
94
+ "step": 2400
95
+ },
96
+ {
97
+ "epoch": 0.832,
98
+ "grad_norm": 0.018439946696162224,
99
+ "learning_rate": 0.012793177895099529,
100
+ "loss": 0.2245,
101
+ "step": 2600
102
+ },
103
+ {
104
+ "epoch": 0.896,
105
+ "grad_norm": 0.017203744500875473,
106
+ "learning_rate": 0.011805335732439208,
107
+ "loss": 0.2212,
108
+ "step": 2800
109
+ },
110
+ {
111
+ "epoch": 0.96,
112
+ "grad_norm": 0.018836447969079018,
113
+ "learning_rate": 0.010798890517852123,
114
+ "loss": 0.2192,
115
+ "step": 3000
116
+ },
117
+ {
118
+ "epoch": 1.024,
119
+ "grad_norm": 0.018993478268384933,
120
+ "learning_rate": 0.009784213149256028,
121
+ "loss": 0.2092,
122
+ "step": 3200
123
+ },
124
+ {
125
+ "epoch": 1.088,
126
+ "grad_norm": 0.02251339517533779,
127
+ "learning_rate": 0.008771759352662289,
128
+ "loss": 0.1962,
129
+ "step": 3400
130
+ },
131
+ {
132
+ "epoch": 1.152,
133
+ "grad_norm": 0.01902582310140133,
134
+ "learning_rate": 0.007771961941321678,
135
+ "loss": 0.1992,
136
+ "step": 3600
137
+ },
138
+ {
139
+ "epoch": 1.216,
140
+ "grad_norm": 0.01809665560722351,
141
+ "learning_rate": 0.006795123310974343,
142
+ "loss": 0.1962,
143
+ "step": 3800
144
+ },
145
+ {
146
+ "epoch": 1.28,
147
+ "grad_norm": 0.016171475872397423,
148
+ "learning_rate": 0.005851309278984825,
149
+ "loss": 0.1932,
150
+ "step": 4000
151
+ },
152
+ {
153
+ "epoch": 1.3439999999999999,
154
+ "grad_norm": 0.017769088968634605,
155
+ "learning_rate": 0.0049502453612950284,
156
+ "loss": 0.1918,
157
+ "step": 4200
158
+ },
159
+ {
160
+ "epoch": 1.408,
161
+ "grad_norm": 0.015108599327504635,
162
+ "learning_rate": 0.004101216556007472,
163
+ "loss": 0.189,
164
+ "step": 4400
165
+ },
166
+ {
167
+ "epoch": 1.472,
168
+ "grad_norm": 0.0172615647315979,
169
+ "learning_rate": 0.0033129716662773125,
170
+ "loss": 0.1888,
171
+ "step": 4600
172
+ },
173
+ {
174
+ "epoch": 1.536,
175
+ "grad_norm": 0.013923043385148048,
176
+ "learning_rate": 0.0025936331484162434,
177
+ "loss": 0.1815,
178
+ "step": 4800
179
+ },
180
+ {
181
+ "epoch": 1.6,
182
+ "grad_norm": 0.014836262911558151,
183
+ "learning_rate": 0.0019506134141769472,
184
+ "loss": 0.1852,
185
+ "step": 5000
186
+ },
187
+ {
188
+ "epoch": 1.6640000000000001,
189
+ "grad_norm": 0.01543043740093708,
190
+ "learning_rate": 0.0013905384496799378,
191
+ "loss": 0.1812,
192
+ "step": 5200
193
+ },
194
+ {
195
+ "epoch": 1.728,
196
+ "grad_norm": 0.015558457002043724,
197
+ "learning_rate": 0.0009191795380501134,
198
+ "loss": 0.1788,
199
+ "step": 5400
200
+ },
201
+ {
202
+ "epoch": 1.792,
203
+ "grad_norm": 0.016567343845963478,
204
+ "learning_rate": 0.0005413937893261456,
205
+ "loss": 0.1768,
206
+ "step": 5600
207
+ },
208
+ {
209
+ "epoch": 1.8559999999999999,
210
+ "grad_norm": 0.013650099746882915,
211
+ "learning_rate": 0.0002610740904511233,
212
+ "loss": 0.1794,
213
+ "step": 5800
214
+ },
215
+ {
216
+ "epoch": 1.92,
217
+ "grad_norm": 0.019280096516013145,
218
+ "learning_rate": 8.110899108401792e-05,
219
+ "loss": 0.1736,
220
+ "step": 6000
221
+ },
222
+ {
223
+ "epoch": 1.984,
224
+ "grad_norm": 0.015858888626098633,
225
+ "learning_rate": 3.3529385879282893e-06,
226
+ "loss": 0.1747,
227
+ "step": 6200
228
+ },
229
+ {
230
+ "epoch": 2.0,
231
+ "step": 6250,
232
+ "total_flos": 4.0647058784256e+18,
233
+ "train_loss": 0.22993893524169923,
234
+ "train_runtime": 5634.1945,
235
+ "train_samples_per_second": 35.498,
236
+ "train_steps_per_second": 1.109
237
+ }
238
+ ],
239
+ "logging_steps": 200,
240
+ "max_steps": 6250,
241
+ "num_input_tokens_seen": 0,
242
+ "num_train_epochs": 2,
243
+ "save_steps": 0,
244
+ "stateful_callbacks": {
245
+ "TrainerControl": {
246
+ "args": {
247
+ "should_epoch_stop": false,
248
+ "should_evaluate": false,
249
+ "should_log": false,
250
+ "should_save": true,
251
+ "should_training_stop": true
252
+ },
253
+ "attributes": {}
254
+ }
255
+ },
256
+ "total_flos": 4.0647058784256e+18,
257
+ "train_batch_size": 32,
258
+ "trial_name": null,
259
+ "trial_params": null
260
+ }
nl_tasks/exp395/run_ex09/ft/adapter_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "T": 1.0,
3
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "inference_mode": false,
6
+ "layers_to_transform": null,
7
+ "modules_to_save": null,
8
+ "num_rotations": 1,
9
+ "peft_type": "ROTATION",
10
+ "r": 16,
11
+ "revision": null,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "target_modules_to_skip": null,
17
+ "task_type": "CAUSAL_LM"
18
+ }
nl_tasks/exp395/run_ex09/ft/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<unk>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
nl_tasks/exp395/run_ex09/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
nl_tasks/exp395/run_ex09/ft/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
nl_tasks/exp395/run_ex09/ft/tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "bos_token": "<s>",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
+ "extra_special_tokens": {},
35
+ "legacy": false,
36
+ "model_max_length": 512,
37
+ "pad_token": "<unk>",
38
+ "padding_side": "right",
39
+ "sp_model_kwargs": {},
40
+ "tokenizer_class": "LlamaTokenizer",
41
+ "unk_token": "<unk>",
42
+ "use_default_system_prompt": false
43
+ }
nl_tasks/exp395/run_ex09/ft2/adapter_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "T": 1.0,
3
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "inference_mode": true,
6
+ "layers_to_transform": null,
7
+ "modules_to_save": null,
8
+ "num_rotations": 1,
9
+ "peft_type": "ROTATION",
10
+ "r": 16,
11
+ "revision": null,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "target_modules_to_skip": null,
17
+ "task_type": "CAUSAL_LM"
18
+ }
nl_tasks/exp395/run_ex09/ft2/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbd1a2c07e6bdebc1f0ef22f1037b61bc06f34f554970fb77b43fe148b47472f
3
+ size 33602915
nl_tasks/exp395/run_ex09/trainer_state.json ADDED
@@ -0,0 +1,509 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 9375,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.064,
14
+ "grad_norm": 0.24448873102664948,
15
+ "learning_rate": 0.00199936844222861,
16
+ "loss": 0.3757,
17
+ "step": 200
18
+ },
19
+ {
20
+ "epoch": 0.128,
21
+ "grad_norm": 0.25495460629463196,
22
+ "learning_rate": 0.0019946753128925747,
23
+ "loss": 0.3045,
24
+ "step": 400
25
+ },
26
+ {
27
+ "epoch": 0.16,
28
+ "eval_loss": 0.2806861102581024,
29
+ "eval_runtime": 18.6896,
30
+ "eval_samples_per_second": 54.094,
31
+ "eval_steps_per_second": 0.856,
32
+ "step": 500
33
+ },
34
+ {
35
+ "epoch": 0.192,
36
+ "grad_norm": 0.12310587614774704,
37
+ "learning_rate": 0.001985425114055429,
38
+ "loss": 0.2727,
39
+ "step": 600
40
+ },
41
+ {
42
+ "epoch": 0.256,
43
+ "grad_norm": 0.15234604477882385,
44
+ "learning_rate": 0.0019716602251735175,
45
+ "loss": 0.2555,
46
+ "step": 800
47
+ },
48
+ {
49
+ "epoch": 0.32,
50
+ "grad_norm": 0.10114556550979614,
51
+ "learning_rate": 0.001953443709594746,
52
+ "loss": 0.2475,
53
+ "step": 1000
54
+ },
55
+ {
56
+ "epoch": 0.32,
57
+ "eval_loss": 0.25135913491249084,
58
+ "eval_runtime": 18.5064,
59
+ "eval_samples_per_second": 54.63,
60
+ "eval_steps_per_second": 0.865,
61
+ "step": 1000
62
+ },
63
+ {
64
+ "epoch": 0.384,
65
+ "grad_norm": 0.09664544463157654,
66
+ "learning_rate": 0.0019308590256360988,
67
+ "loss": 0.2449,
68
+ "step": 1200
69
+ },
70
+ {
71
+ "epoch": 0.448,
72
+ "grad_norm": 0.09191286563873291,
73
+ "learning_rate": 0.0019040096442223358,
74
+ "loss": 0.2387,
75
+ "step": 1400
76
+ },
77
+ {
78
+ "epoch": 0.48,
79
+ "eval_loss": 0.24170711636543274,
80
+ "eval_runtime": 18.503,
81
+ "eval_samples_per_second": 54.64,
82
+ "eval_steps_per_second": 0.865,
83
+ "step": 1500
84
+ },
85
+ {
86
+ "epoch": 0.512,
87
+ "grad_norm": 0.08609094470739365,
88
+ "learning_rate": 0.0018730185748376353,
89
+ "loss": 0.2332,
90
+ "step": 1600
91
+ },
92
+ {
93
+ "epoch": 0.576,
94
+ "grad_norm": 0.08084086328744888,
95
+ "learning_rate": 0.0018380278019620281,
96
+ "loss": 0.2295,
97
+ "step": 1800
98
+ },
99
+ {
100
+ "epoch": 0.64,
101
+ "grad_norm": 0.08400174230337143,
102
+ "learning_rate": 0.0017991976345745596,
103
+ "loss": 0.2278,
104
+ "step": 2000
105
+ },
106
+ {
107
+ "epoch": 0.64,
108
+ "eval_loss": 0.22995951771736145,
109
+ "eval_runtime": 18.5154,
110
+ "eval_samples_per_second": 54.603,
111
+ "eval_steps_per_second": 0.864,
112
+ "step": 2000
113
+ },
114
+ {
115
+ "epoch": 0.704,
116
+ "grad_norm": 0.06514922529459,
117
+ "learning_rate": 0.0017567059717034153,
118
+ "loss": 0.2236,
119
+ "step": 2200
120
+ },
121
+ {
122
+ "epoch": 0.768,
123
+ "grad_norm": 0.061288416385650635,
124
+ "learning_rate": 0.0017107474873878648,
125
+ "loss": 0.2217,
126
+ "step": 2400
127
+ },
128
+ {
129
+ "epoch": 0.8,
130
+ "eval_loss": 0.2240372747182846,
131
+ "eval_runtime": 18.5034,
132
+ "eval_samples_per_second": 54.639,
133
+ "eval_steps_per_second": 0.865,
134
+ "step": 2500
135
+ },
136
+ {
137
+ "epoch": 0.832,
138
+ "grad_norm": 0.0703856498003006,
139
+ "learning_rate": 0.0016615327387860851,
140
+ "loss": 0.2195,
141
+ "step": 2600
142
+ },
143
+ {
144
+ "epoch": 0.896,
145
+ "grad_norm": 0.06190377473831177,
146
+ "learning_rate": 0.001609287201515044,
147
+ "loss": 0.2192,
148
+ "step": 2800
149
+ },
150
+ {
151
+ "epoch": 0.96,
152
+ "grad_norm": 0.058891598135232925,
153
+ "learning_rate": 0.0015542502366419986,
154
+ "loss": 0.2119,
155
+ "step": 3000
156
+ },
157
+ {
158
+ "epoch": 0.96,
159
+ "eval_loss": 0.21980270743370056,
160
+ "eval_runtime": 18.5017,
161
+ "eval_samples_per_second": 54.644,
162
+ "eval_steps_per_second": 0.865,
163
+ "step": 3000
164
+ },
165
+ {
166
+ "epoch": 1.024,
167
+ "grad_norm": 0.08175381273031235,
168
+ "learning_rate": 0.001496673994060291,
169
+ "loss": 0.2067,
170
+ "step": 3200
171
+ },
172
+ {
173
+ "epoch": 1.088,
174
+ "grad_norm": 0.058054033666849136,
175
+ "learning_rate": 0.0014368222572736001,
176
+ "loss": 0.1904,
177
+ "step": 3400
178
+ },
179
+ {
180
+ "epoch": 1.12,
181
+ "eval_loss": 0.21591630578041077,
182
+ "eval_runtime": 18.5104,
183
+ "eval_samples_per_second": 54.618,
184
+ "eval_steps_per_second": 0.864,
185
+ "step": 3500
186
+ },
187
+ {
188
+ "epoch": 1.152,
189
+ "grad_norm": 0.06751835346221924,
190
+ "learning_rate": 0.0013749692348812186,
191
+ "loss": 0.1906,
192
+ "step": 3600
193
+ },
194
+ {
195
+ "epoch": 1.216,
196
+ "grad_norm": 0.06137789785861969,
197
+ "learning_rate": 0.0013113983043011267,
198
+ "loss": 0.1906,
199
+ "step": 3800
200
+ },
201
+ {
202
+ "epoch": 1.28,
203
+ "grad_norm": 0.05203723534941673,
204
+ "learning_rate": 0.0012464007134864482,
205
+ "loss": 0.1883,
206
+ "step": 4000
207
+ },
208
+ {
209
+ "epoch": 1.28,
210
+ "eval_loss": 0.21353089809417725,
211
+ "eval_runtime": 18.5177,
212
+ "eval_samples_per_second": 54.596,
213
+ "eval_steps_per_second": 0.864,
214
+ "step": 4000
215
+ },
216
+ {
217
+ "epoch": 1.3439999999999999,
218
+ "grad_norm": 0.0589132197201252,
219
+ "learning_rate": 0.0011802742465833352,
220
+ "loss": 0.1902,
221
+ "step": 4200
222
+ },
223
+ {
224
+ "epoch": 1.408,
225
+ "grad_norm": 0.06806980073451996,
226
+ "learning_rate": 0.0011133218596435107,
227
+ "loss": 0.1883,
228
+ "step": 4400
229
+ },
230
+ {
231
+ "epoch": 1.44,
232
+ "eval_loss": 0.20915259420871735,
233
+ "eval_runtime": 18.5243,
234
+ "eval_samples_per_second": 54.577,
235
+ "eval_steps_per_second": 0.864,
236
+ "step": 4500
237
+ },
238
+ {
239
+ "epoch": 1.472,
240
+ "grad_norm": 0.06540437787771225,
241
+ "learning_rate": 0.0010458502926419136,
242
+ "loss": 0.189,
243
+ "step": 4600
244
+ },
245
+ {
246
+ "epoch": 1.536,
247
+ "grad_norm": 0.05099968612194061,
248
+ "learning_rate": 0.0009781686641584368,
249
+ "loss": 0.1862,
250
+ "step": 4800
251
+ },
252
+ {
253
+ "epoch": 1.6,
254
+ "grad_norm": 0.056332822889089584,
255
+ "learning_rate": 0.0009105870551621664,
256
+ "loss": 0.184,
257
+ "step": 5000
258
+ },
259
+ {
260
+ "epoch": 1.6,
261
+ "eval_loss": 0.20469947159290314,
262
+ "eval_runtime": 18.5065,
263
+ "eval_samples_per_second": 54.629,
264
+ "eval_steps_per_second": 0.865,
265
+ "step": 5000
266
+ },
267
+ {
268
+ "epoch": 1.6640000000000001,
269
+ "grad_norm": 0.04745101556181908,
270
+ "learning_rate": 0.0008434150883864713,
271
+ "loss": 0.1866,
272
+ "step": 5200
273
+ },
274
+ {
275
+ "epoch": 1.728,
276
+ "grad_norm": 0.04577454924583435,
277
+ "learning_rate": 0.0007769605098034851,
278
+ "loss": 0.186,
279
+ "step": 5400
280
+ },
281
+ {
282
+ "epoch": 1.76,
283
+ "eval_loss": 0.20063205063343048,
284
+ "eval_runtime": 18.5081,
285
+ "eval_samples_per_second": 54.625,
286
+ "eval_steps_per_second": 0.864,
287
+ "step": 5500
288
+ },
289
+ {
290
+ "epoch": 1.792,
291
+ "grad_norm": 0.04938381910324097,
292
+ "learning_rate": 0.0007115277786969001,
293
+ "loss": 0.1831,
294
+ "step": 5600
295
+ },
296
+ {
297
+ "epoch": 1.8559999999999999,
298
+ "grad_norm": 0.05956464633345604,
299
+ "learning_rate": 0.0006474166727926085,
300
+ "loss": 0.1804,
301
+ "step": 5800
302
+ },
303
+ {
304
+ "epoch": 1.92,
305
+ "grad_norm": 0.050873901695013046,
306
+ "learning_rate": 0.0005849209148377339,
307
+ "loss": 0.1808,
308
+ "step": 6000
309
+ },
310
+ {
311
+ "epoch": 1.92,
312
+ "eval_loss": 0.19810360670089722,
313
+ "eval_runtime": 22.7592,
314
+ "eval_samples_per_second": 44.422,
315
+ "eval_steps_per_second": 0.703,
316
+ "step": 6000
317
+ },
318
+ {
319
+ "epoch": 1.984,
320
+ "grad_norm": 0.09603903442621231,
321
+ "learning_rate": 0.0005243268269203323,
322
+ "loss": 0.1806,
323
+ "step": 6200
324
+ },
325
+ {
326
+ "epoch": 2.048,
327
+ "grad_norm": 0.05339692533016205,
328
+ "learning_rate": 0.0004659120186949479,
329
+ "loss": 0.1613,
330
+ "step": 6400
331
+ },
332
+ {
333
+ "epoch": 2.08,
334
+ "eval_loss": 0.1972031444311142,
335
+ "eval_runtime": 18.5168,
336
+ "eval_samples_per_second": 54.599,
337
+ "eval_steps_per_second": 0.864,
338
+ "step": 6500
339
+ },
340
+ {
341
+ "epoch": 2.112,
342
+ "grad_norm": 0.053911786526441574,
343
+ "learning_rate": 0.0004099441155238748,
344
+ "loss": 0.1544,
345
+ "step": 6600
346
+ },
347
+ {
348
+ "epoch": 2.176,
349
+ "grad_norm": 0.050456177443265915,
350
+ "learning_rate": 0.0003566795323610935,
351
+ "loss": 0.155,
352
+ "step": 6800
353
+ },
354
+ {
355
+ "epoch": 2.24,
356
+ "grad_norm": 0.05330171436071396,
357
+ "learning_rate": 0.0003063622989963002,
358
+ "loss": 0.1529,
359
+ "step": 7000
360
+ },
361
+ {
362
+ "epoch": 2.24,
363
+ "eval_loss": 0.1969524472951889,
364
+ "eval_runtime": 18.4963,
365
+ "eval_samples_per_second": 54.659,
366
+ "eval_steps_per_second": 0.865,
367
+ "step": 7000
368
+ },
369
+ {
370
+ "epoch": 2.304,
371
+ "grad_norm": 0.053924355655908585,
372
+ "learning_rate": 0.0002592229420411184,
373
+ "loss": 0.1545,
374
+ "step": 7200
375
+ },
376
+ {
377
+ "epoch": 2.368,
378
+ "grad_norm": 0.055782243609428406,
379
+ "learning_rate": 0.00021547742877964395,
380
+ "loss": 0.1569,
381
+ "step": 7400
382
+ },
383
+ {
384
+ "epoch": 2.4,
385
+ "eval_loss": 0.19416877627372742,
386
+ "eval_runtime": 18.4948,
387
+ "eval_samples_per_second": 54.664,
388
+ "eval_steps_per_second": 0.865,
389
+ "step": 7500
390
+ },
391
+ {
392
+ "epoch": 2.432,
393
+ "grad_norm": 0.05896177887916565,
394
+ "learning_rate": 0.00017532617772202807,
395
+ "loss": 0.1557,
396
+ "step": 7600
397
+ },
398
+ {
399
+ "epoch": 2.496,
400
+ "grad_norm": 0.05818909779191017,
401
+ "learning_rate": 0.0001389531403942159,
402
+ "loss": 0.1558,
403
+ "step": 7800
404
+ },
405
+ {
406
+ "epoch": 2.56,
407
+ "grad_norm": 0.05467315390706062,
408
+ "learning_rate": 0.00010652495857058375,
409
+ "loss": 0.1535,
410
+ "step": 8000
411
+ },
412
+ {
413
+ "epoch": 2.56,
414
+ "eval_loss": 0.19276997447013855,
415
+ "eval_runtime": 18.5042,
416
+ "eval_samples_per_second": 54.636,
417
+ "eval_steps_per_second": 0.865,
418
+ "step": 8000
419
+ },
420
+ {
421
+ "epoch": 2.624,
422
+ "grad_norm": 0.052459824830293655,
423
+ "learning_rate": 7.81902008105898e-05,
424
+ "loss": 0.1515,
425
+ "step": 8200
426
+ },
427
+ {
428
+ "epoch": 2.6879999999999997,
429
+ "grad_norm": 0.05594847351312637,
430
+ "learning_rate": 5.407868179721254e-05,
431
+ "loss": 0.1535,
432
+ "step": 8400
433
+ },
434
+ {
435
+ "epoch": 2.7199999999999998,
436
+ "eval_loss": 0.1920287013053894,
437
+ "eval_runtime": 18.5152,
438
+ "eval_samples_per_second": 54.604,
439
+ "eval_steps_per_second": 0.864,
440
+ "step": 8500
441
+ },
442
+ {
443
+ "epoch": 2.752,
444
+ "grad_norm": 0.054888200014829636,
445
+ "learning_rate": 3.430086759560602e-05,
446
+ "loss": 0.1503,
447
+ "step": 8600
448
+ },
449
+ {
450
+ "epoch": 2.816,
451
+ "grad_norm": 0.06011143699288368,
452
+ "learning_rate": 1.8947369556759353e-05,
453
+ "loss": 0.1535,
454
+ "step": 8800
455
+ },
456
+ {
457
+ "epoch": 2.88,
458
+ "grad_norm": 0.04984534904360771,
459
+ "learning_rate": 8.088529184821281e-06,
460
+ "loss": 0.1522,
461
+ "step": 9000
462
+ },
463
+ {
464
+ "epoch": 2.88,
465
+ "eval_loss": 0.19162563979625702,
466
+ "eval_runtime": 18.5125,
467
+ "eval_samples_per_second": 54.612,
468
+ "eval_steps_per_second": 0.864,
469
+ "step": 9000
470
+ },
471
+ {
472
+ "epoch": 2.944,
473
+ "grad_norm": 0.05723896995186806,
474
+ "learning_rate": 1.774095870004544e-06,
475
+ "loss": 0.1492,
476
+ "step": 9200
477
+ },
478
+ {
479
+ "epoch": 3.0,
480
+ "step": 9375,
481
+ "total_flos": 6.0970588176384e+18,
482
+ "train_loss": 0.19560041158040364,
483
+ "train_runtime": 8921.5527,
484
+ "train_samples_per_second": 33.626,
485
+ "train_steps_per_second": 1.051
486
+ }
487
+ ],
488
+ "logging_steps": 200,
489
+ "max_steps": 9375,
490
+ "num_input_tokens_seen": 0,
491
+ "num_train_epochs": 3,
492
+ "save_steps": 0,
493
+ "stateful_callbacks": {
494
+ "TrainerControl": {
495
+ "args": {
496
+ "should_epoch_stop": false,
497
+ "should_evaluate": false,
498
+ "should_log": false,
499
+ "should_save": true,
500
+ "should_training_stop": true
501
+ },
502
+ "attributes": {}
503
+ }
504
+ },
505
+ "total_flos": 6.0970588176384e+18,
506
+ "train_batch_size": 32,
507
+ "trial_name": null,
508
+ "trial_params": null
509
+ }
nl_tasks/exp395/run_ex10/ft/adapter_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "T": 1.0,
3
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "inference_mode": false,
6
+ "layers_to_transform": null,
7
+ "modules_to_save": null,
8
+ "num_rotations": 1,
9
+ "peft_type": "ROTATION",
10
+ "r": 16,
11
+ "revision": null,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "target_modules_to_skip": null,
17
+ "task_type": "CAUSAL_LM"
18
+ }
nl_tasks/exp395/run_ex10/ft/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<unk>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
nl_tasks/exp395/run_ex10/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
nl_tasks/exp395/run_ex10/ft/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
nl_tasks/exp395/run_ex10/ft/tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "bos_token": "<s>",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
+ "extra_special_tokens": {},
35
+ "legacy": false,
36
+ "model_max_length": 512,
37
+ "pad_token": "<unk>",
38
+ "padding_side": "right",
39
+ "sp_model_kwargs": {},
40
+ "tokenizer_class": "LlamaTokenizer",
41
+ "unk_token": "<unk>",
42
+ "use_default_system_prompt": false
43
+ }
nl_tasks/exp395/run_ex10/ft2/adapter_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "T": 1.0,
3
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "inference_mode": true,
6
+ "layers_to_transform": null,
7
+ "modules_to_save": null,
8
+ "num_rotations": 1,
9
+ "peft_type": "ROTATION",
10
+ "r": 16,
11
+ "revision": null,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "target_modules_to_skip": null,
17
+ "task_type": "CAUSAL_LM"
18
+ }
nl_tasks/exp395/run_ex10/ft2/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55f078d358c3a5c84a88c3b5fc6605b32726f64997693d45b47f259deaa97c3e
3
+ size 33602915
nl_tasks/exp395/run_ex10/trainer_state.json ADDED
@@ -0,0 +1,356 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 6250,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.064,
14
+ "grad_norm": 0.2293911576271057,
15
+ "learning_rate": 0.001997616502308297,
16
+ "loss": 0.4044,
17
+ "step": 200
18
+ },
19
+ {
20
+ "epoch": 0.128,
21
+ "grad_norm": 0.18487609922885895,
22
+ "learning_rate": 0.00198548107876105,
23
+ "loss": 0.3009,
24
+ "step": 400
25
+ },
26
+ {
27
+ "epoch": 0.16,
28
+ "eval_loss": 0.2799297869205475,
29
+ "eval_runtime": 18.7558,
30
+ "eval_samples_per_second": 53.903,
31
+ "eval_steps_per_second": 0.853,
32
+ "step": 500
33
+ },
34
+ {
35
+ "epoch": 0.192,
36
+ "grad_norm": 0.13151583075523376,
37
+ "learning_rate": 0.001963190781883806,
38
+ "loss": 0.2711,
39
+ "step": 600
40
+ },
41
+ {
42
+ "epoch": 0.256,
43
+ "grad_norm": 0.13094772398471832,
44
+ "learning_rate": 0.0019309753016687476,
45
+ "loss": 0.2554,
46
+ "step": 800
47
+ },
48
+ {
49
+ "epoch": 0.32,
50
+ "grad_norm": 0.09965614229440689,
51
+ "learning_rate": 0.0018891666019941258,
52
+ "loss": 0.2469,
53
+ "step": 1000
54
+ },
55
+ {
56
+ "epoch": 0.32,
57
+ "eval_loss": 0.25120654702186584,
58
+ "eval_runtime": 18.4999,
59
+ "eval_samples_per_second": 54.649,
60
+ "eval_steps_per_second": 0.865,
61
+ "step": 1000
62
+ },
63
+ {
64
+ "epoch": 0.384,
65
+ "grad_norm": 0.10732943564653397,
66
+ "learning_rate": 0.001838195499908018,
67
+ "loss": 0.2444,
68
+ "step": 1200
69
+ },
70
+ {
71
+ "epoch": 0.448,
72
+ "grad_norm": 0.09097783267498016,
73
+ "learning_rate": 0.001778587226281244,
74
+ "loss": 0.2377,
75
+ "step": 1400
76
+ },
77
+ {
78
+ "epoch": 0.48,
79
+ "eval_loss": 0.23919327557086945,
80
+ "eval_runtime": 18.4946,
81
+ "eval_samples_per_second": 54.665,
82
+ "eval_steps_per_second": 0.865,
83
+ "step": 1500
84
+ },
85
+ {
86
+ "epoch": 0.512,
87
+ "grad_norm": 0.08693786710500717,
88
+ "learning_rate": 0.0017109560135746254,
89
+ "loss": 0.232,
90
+ "step": 1600
91
+ },
92
+ {
93
+ "epoch": 0.576,
94
+ "grad_norm": 0.0780855268239975,
95
+ "learning_rate": 0.001635998766490841,
96
+ "loss": 0.2284,
97
+ "step": 1800
98
+ },
99
+ {
100
+ "epoch": 0.64,
101
+ "grad_norm": 0.0817127600312233,
102
+ "learning_rate": 0.0015544878807315496,
103
+ "loss": 0.2258,
104
+ "step": 2000
105
+ },
106
+ {
107
+ "epoch": 0.64,
108
+ "eval_loss": 0.2277495265007019,
109
+ "eval_runtime": 18.487,
110
+ "eval_samples_per_second": 54.687,
111
+ "eval_steps_per_second": 0.865,
112
+ "step": 2000
113
+ },
114
+ {
115
+ "epoch": 0.704,
116
+ "grad_norm": 0.0686568021774292,
117
+ "learning_rate": 0.0014672632838587717,
118
+ "loss": 0.2217,
119
+ "step": 2200
120
+ },
121
+ {
122
+ "epoch": 0.768,
123
+ "grad_norm": 0.0656941682100296,
124
+ "learning_rate": 0.0013752237802753448,
125
+ "loss": 0.2195,
126
+ "step": 2400
127
+ },
128
+ {
129
+ "epoch": 0.8,
130
+ "eval_loss": 0.2217738926410675,
131
+ "eval_runtime": 18.5043,
132
+ "eval_samples_per_second": 54.636,
133
+ "eval_steps_per_second": 0.865,
134
+ "step": 2500
135
+ },
136
+ {
137
+ "epoch": 0.832,
138
+ "grad_norm": 0.0752493143081665,
139
+ "learning_rate": 0.0012793177895099527,
140
+ "loss": 0.2167,
141
+ "step": 2600
142
+ },
143
+ {
144
+ "epoch": 0.896,
145
+ "grad_norm": 0.08712925016880035,
146
+ "learning_rate": 0.0011805335732439208,
147
+ "loss": 0.2158,
148
+ "step": 2800
149
+ },
150
+ {
151
+ "epoch": 0.96,
152
+ "grad_norm": 0.05949317663908005,
153
+ "learning_rate": 0.0010798890517852123,
154
+ "loss": 0.2085,
155
+ "step": 3000
156
+ },
157
+ {
158
+ "epoch": 0.96,
159
+ "eval_loss": 0.21555665135383606,
160
+ "eval_runtime": 18.496,
161
+ "eval_samples_per_second": 54.66,
162
+ "eval_steps_per_second": 0.865,
163
+ "step": 3000
164
+ },
165
+ {
166
+ "epoch": 1.024,
167
+ "grad_norm": 0.07340498268604279,
168
+ "learning_rate": 0.0009784213149256027,
169
+ "loss": 0.2019,
170
+ "step": 3200
171
+ },
172
+ {
173
+ "epoch": 1.088,
174
+ "grad_norm": 0.06179428473114967,
175
+ "learning_rate": 0.0008771759352662288,
176
+ "loss": 0.184,
177
+ "step": 3400
178
+ },
179
+ {
180
+ "epoch": 1.12,
181
+ "eval_loss": 0.21221420168876648,
182
+ "eval_runtime": 18.4908,
183
+ "eval_samples_per_second": 54.676,
184
+ "eval_steps_per_second": 0.865,
185
+ "step": 3500
186
+ },
187
+ {
188
+ "epoch": 1.152,
189
+ "grad_norm": 0.07018924504518509,
190
+ "learning_rate": 0.0007771961941321678,
191
+ "loss": 0.1836,
192
+ "step": 3600
193
+ },
194
+ {
195
+ "epoch": 1.216,
196
+ "grad_norm": 0.07013561576604843,
197
+ "learning_rate": 0.0006795123310974343,
198
+ "loss": 0.1827,
199
+ "step": 3800
200
+ },
201
+ {
202
+ "epoch": 1.28,
203
+ "grad_norm": 0.061091359704732895,
204
+ "learning_rate": 0.0005851309278984824,
205
+ "loss": 0.1798,
206
+ "step": 4000
207
+ },
208
+ {
209
+ "epoch": 1.28,
210
+ "eval_loss": 0.20682914555072784,
211
+ "eval_runtime": 18.5014,
212
+ "eval_samples_per_second": 54.645,
213
+ "eval_steps_per_second": 0.865,
214
+ "step": 4000
215
+ },
216
+ {
217
+ "epoch": 1.3439999999999999,
218
+ "grad_norm": 0.0593683086335659,
219
+ "learning_rate": 0.0004950245361295028,
220
+ "loss": 0.1807,
221
+ "step": 4200
222
+ },
223
+ {
224
+ "epoch": 1.408,
225
+ "grad_norm": 0.08087774366140366,
226
+ "learning_rate": 0.0004101216556007472,
227
+ "loss": 0.1788,
228
+ "step": 4400
229
+ },
230
+ {
231
+ "epoch": 1.44,
232
+ "eval_loss": 0.2025197595357895,
233
+ "eval_runtime": 18.5002,
234
+ "eval_samples_per_second": 54.648,
235
+ "eval_steps_per_second": 0.865,
236
+ "step": 4500
237
+ },
238
+ {
239
+ "epoch": 1.472,
240
+ "grad_norm": 0.06437215209007263,
241
+ "learning_rate": 0.00033129716662773124,
242
+ "loss": 0.1788,
243
+ "step": 4600
244
+ },
245
+ {
246
+ "epoch": 1.536,
247
+ "grad_norm": 0.06128371134400368,
248
+ "learning_rate": 0.0002593633148416243,
249
+ "loss": 0.1757,
250
+ "step": 4800
251
+ },
252
+ {
253
+ "epoch": 1.6,
254
+ "grad_norm": 0.061387427151203156,
255
+ "learning_rate": 0.00019506134141769472,
256
+ "loss": 0.1731,
257
+ "step": 5000
258
+ },
259
+ {
260
+ "epoch": 1.6,
261
+ "eval_loss": 0.198740616440773,
262
+ "eval_runtime": 18.5095,
263
+ "eval_samples_per_second": 54.621,
264
+ "eval_steps_per_second": 0.864,
265
+ "step": 5000
266
+ },
267
+ {
268
+ "epoch": 1.6640000000000001,
269
+ "grad_norm": 0.060046833008527756,
270
+ "learning_rate": 0.00013905384496799378,
271
+ "loss": 0.1756,
272
+ "step": 5200
273
+ },
274
+ {
275
+ "epoch": 1.728,
276
+ "grad_norm": 0.05966249108314514,
277
+ "learning_rate": 9.191795380501134e-05,
278
+ "loss": 0.1752,
279
+ "step": 5400
280
+ },
281
+ {
282
+ "epoch": 1.76,
283
+ "eval_loss": 0.1962076872587204,
284
+ "eval_runtime": 18.494,
285
+ "eval_samples_per_second": 54.667,
286
+ "eval_steps_per_second": 0.865,
287
+ "step": 5500
288
+ },
289
+ {
290
+ "epoch": 1.792,
291
+ "grad_norm": 0.06310608237981796,
292
+ "learning_rate": 5.413937893261456e-05,
293
+ "loss": 0.1732,
294
+ "step": 5600
295
+ },
296
+ {
297
+ "epoch": 1.8559999999999999,
298
+ "grad_norm": 0.06762094050645828,
299
+ "learning_rate": 2.610740904511233e-05,
300
+ "loss": 0.171,
301
+ "step": 5800
302
+ },
303
+ {
304
+ "epoch": 1.92,
305
+ "grad_norm": 0.06283057481050491,
306
+ "learning_rate": 8.110899108401792e-06,
307
+ "loss": 0.1717,
308
+ "step": 6000
309
+ },
310
+ {
311
+ "epoch": 1.92,
312
+ "eval_loss": 0.1952924132347107,
313
+ "eval_runtime": 18.4907,
314
+ "eval_samples_per_second": 54.676,
315
+ "eval_steps_per_second": 0.865,
316
+ "step": 6000
317
+ },
318
+ {
319
+ "epoch": 1.984,
320
+ "grad_norm": 0.06748773157596588,
321
+ "learning_rate": 3.352938587928289e-07,
322
+ "loss": 0.1728,
323
+ "step": 6200
324
+ },
325
+ {
326
+ "epoch": 2.0,
327
+ "step": 6250,
328
+ "total_flos": 4.0647058784256e+18,
329
+ "train_loss": 0.2122136308288574,
330
+ "train_runtime": 5832.5182,
331
+ "train_samples_per_second": 34.291,
332
+ "train_steps_per_second": 1.072
333
+ }
334
+ ],
335
+ "logging_steps": 200,
336
+ "max_steps": 6250,
337
+ "num_input_tokens_seen": 0,
338
+ "num_train_epochs": 2,
339
+ "save_steps": 0,
340
+ "stateful_callbacks": {
341
+ "TrainerControl": {
342
+ "args": {
343
+ "should_epoch_stop": false,
344
+ "should_evaluate": false,
345
+ "should_log": false,
346
+ "should_save": true,
347
+ "should_training_stop": true
348
+ },
349
+ "attributes": {}
350
+ }
351
+ },
352
+ "total_flos": 4.0647058784256e+18,
353
+ "train_batch_size": 32,
354
+ "trial_name": null,
355
+ "trial_params": null
356
+ }
nl_tasks/exp_init/run_ex01/ft/adapter_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "T": 1.0,
3
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "inference_mode": false,
6
+ "layers_to_transform": null,
7
+ "modules_to_save": null,
8
+ "num_rotations": 1,
9
+ "peft_type": "ROTATION",
10
+ "r": 16,
11
+ "revision": null,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "target_modules_to_skip": null,
17
+ "task_type": "CAUSAL_LM"
18
+ }
nl_tasks/exp_init/run_ex01/ft/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[PAD]": 32000
3
+ }
nl_tasks/exp_init/run_ex01/ft/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "</s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "</s>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
nl_tasks/exp_init/run_ex01/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
nl_tasks/exp_init/run_ex01/ft/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
nl_tasks/exp_init/run_ex01/ft/tokenizer_config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "[PAD]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ }
38
+ },
39
+ "bos_token": "</s>",
40
+ "clean_up_tokenization_spaces": false,
41
+ "eos_token": "</s>",
42
+ "extra_special_tokens": {},
43
+ "legacy": false,
44
+ "model_max_length": 512,
45
+ "pad_token": "[PAD]",
46
+ "padding_side": "right",
47
+ "sp_model_kwargs": {},
48
+ "tokenizer_class": "LlamaTokenizer",
49
+ "unk_token": "</s>",
50
+ "use_default_system_prompt": false
51
+ }
nl_tasks/exp_init/run_ex01/ft2/adapter_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "T": 1.0,
3
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "inference_mode": true,
6
+ "layers_to_transform": null,
7
+ "modules_to_save": null,
8
+ "num_rotations": 1,
9
+ "peft_type": "ROTATION",
10
+ "r": 16,
11
+ "revision": null,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "target_modules_to_skip": null,
17
+ "task_type": "CAUSAL_LM"
18
+ }
nl_tasks/exp_init/run_ex01/ft2/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17cee51a669e30c3b89f014d5339b6df2b9f4008290cbd4f6b5293d21a012450
3
+ size 33602915
nl_tasks/exp_init/run_ex01/trainer_state.json ADDED
@@ -0,0 +1,928 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.0,
6
+ "eval_steps": 100,
7
+ "global_step": 2494,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.020048115477145148,
14
+ "grad_norm": 0.3128751516342163,
15
+ "learning_rate": 9.6e-05,
16
+ "loss": 0.6108,
17
+ "step": 25
18
+ },
19
+ {
20
+ "epoch": 0.040096230954290296,
21
+ "grad_norm": 0.2898954153060913,
22
+ "learning_rate": 0.00019600000000000002,
23
+ "loss": 0.3986,
24
+ "step": 50
25
+ },
26
+ {
27
+ "epoch": 0.060144346431435444,
28
+ "grad_norm": 0.2638753652572632,
29
+ "learning_rate": 0.000296,
30
+ "loss": 0.3553,
31
+ "step": 75
32
+ },
33
+ {
34
+ "epoch": 0.08019246190858059,
35
+ "grad_norm": 0.2679823935031891,
36
+ "learning_rate": 0.00039600000000000003,
37
+ "loss": 0.3276,
38
+ "step": 100
39
+ },
40
+ {
41
+ "epoch": 0.08019246190858059,
42
+ "eval_loss": 0.3061896860599518,
43
+ "eval_runtime": 2.2347,
44
+ "eval_samples_per_second": 46.538,
45
+ "eval_steps_per_second": 0.895,
46
+ "step": 100
47
+ },
48
+ {
49
+ "epoch": 0.10024057738572574,
50
+ "grad_norm": 0.3121950924396515,
51
+ "learning_rate": 0.000496,
52
+ "loss": 0.3272,
53
+ "step": 125
54
+ },
55
+ {
56
+ "epoch": 0.12028869286287089,
57
+ "grad_norm": 0.2655491828918457,
58
+ "learning_rate": 0.000596,
59
+ "loss": 0.3186,
60
+ "step": 150
61
+ },
62
+ {
63
+ "epoch": 0.14033680834001605,
64
+ "grad_norm": 0.30342063307762146,
65
+ "learning_rate": 0.000696,
66
+ "loss": 0.3094,
67
+ "step": 175
68
+ },
69
+ {
70
+ "epoch": 0.16038492381716118,
71
+ "grad_norm": 0.3187066614627838,
72
+ "learning_rate": 0.000796,
73
+ "loss": 0.3092,
74
+ "step": 200
75
+ },
76
+ {
77
+ "epoch": 0.16038492381716118,
78
+ "eval_loss": 0.29791951179504395,
79
+ "eval_runtime": 2.0135,
80
+ "eval_samples_per_second": 51.651,
81
+ "eval_steps_per_second": 0.993,
82
+ "step": 200
83
+ },
84
+ {
85
+ "epoch": 0.18043303929430635,
86
+ "grad_norm": 0.29701462388038635,
87
+ "learning_rate": 0.000896,
88
+ "loss": 0.303,
89
+ "step": 225
90
+ },
91
+ {
92
+ "epoch": 0.20048115477145148,
93
+ "grad_norm": 0.3302502930164337,
94
+ "learning_rate": 0.000996,
95
+ "loss": 0.302,
96
+ "step": 250
97
+ },
98
+ {
99
+ "epoch": 0.22052927024859664,
100
+ "grad_norm": 0.2812274694442749,
101
+ "learning_rate": 0.0009997177878718869,
102
+ "loss": 0.6194,
103
+ "step": 275
104
+ },
105
+ {
106
+ "epoch": 0.24057738572574178,
107
+ "grad_norm": 0.2586809992790222,
108
+ "learning_rate": 0.0009988239768018291,
109
+ "loss": 0.3014,
110
+ "step": 300
111
+ },
112
+ {
113
+ "epoch": 0.24057738572574178,
114
+ "eval_loss": 0.2832469344139099,
115
+ "eval_runtime": 2.0065,
116
+ "eval_samples_per_second": 51.832,
117
+ "eval_steps_per_second": 0.997,
118
+ "step": 300
119
+ },
120
+ {
121
+ "epoch": 0.2606255012028869,
122
+ "grad_norm": 0.2703372538089752,
123
+ "learning_rate": 0.0009973191715938715,
124
+ "loss": 0.2981,
125
+ "step": 325
126
+ },
127
+ {
128
+ "epoch": 0.2806736166800321,
129
+ "grad_norm": 0.2600429356098175,
130
+ "learning_rate": 0.0009952052154376025,
131
+ "loss": 0.2955,
132
+ "step": 350
133
+ },
134
+ {
135
+ "epoch": 0.30072173215717724,
136
+ "grad_norm": 0.21045513451099396,
137
+ "learning_rate": 0.0009924846976528616,
138
+ "loss": 0.2865,
139
+ "step": 375
140
+ },
141
+ {
142
+ "epoch": 0.32076984763432237,
143
+ "grad_norm": 0.21870078146457672,
144
+ "learning_rate": 0.0009891609505181592,
145
+ "loss": 0.2754,
146
+ "step": 400
147
+ },
148
+ {
149
+ "epoch": 0.32076984763432237,
150
+ "eval_loss": 0.2693285048007965,
151
+ "eval_runtime": 2.008,
152
+ "eval_samples_per_second": 51.792,
153
+ "eval_steps_per_second": 0.996,
154
+ "step": 400
155
+ },
156
+ {
157
+ "epoch": 0.3408179631114675,
158
+ "grad_norm": 0.23548808693885803,
159
+ "learning_rate": 0.0009852380451890721,
160
+ "loss": 0.2812,
161
+ "step": 425
162
+ },
163
+ {
164
+ "epoch": 0.3608660785886127,
165
+ "grad_norm": 0.22311964631080627,
166
+ "learning_rate": 0.0009807207867116115,
167
+ "loss": 0.2868,
168
+ "step": 450
169
+ },
170
+ {
171
+ "epoch": 0.3809141940657578,
172
+ "grad_norm": 0.21461476385593414,
173
+ "learning_rate": 0.0009756147081366672,
174
+ "loss": 0.2765,
175
+ "step": 475
176
+ },
177
+ {
178
+ "epoch": 0.40096230954290296,
179
+ "grad_norm": 0.19619832932949066,
180
+ "learning_rate": 0.0009699260637427467,
181
+ "loss": 0.2759,
182
+ "step": 500
183
+ },
184
+ {
185
+ "epoch": 0.40096230954290296,
186
+ "eval_loss": 0.25922319293022156,
187
+ "eval_runtime": 2.0086,
188
+ "eval_samples_per_second": 51.777,
189
+ "eval_steps_per_second": 0.996,
190
+ "step": 500
191
+ },
192
+ {
193
+ "epoch": 0.4210104250200481,
194
+ "grad_norm": 0.16149669885635376,
195
+ "learning_rate": 0.0009636618213753006,
196
+ "loss": 0.2731,
197
+ "step": 525
198
+ },
199
+ {
200
+ "epoch": 0.4410585404971933,
201
+ "grad_norm": 0.20750294625759125,
202
+ "learning_rate": 0.0009568296539120225,
203
+ "loss": 0.2759,
204
+ "step": 550
205
+ },
206
+ {
207
+ "epoch": 0.4611066559743384,
208
+ "grad_norm": 0.18264305591583252,
209
+ "learning_rate": 0.0009494379298645788,
210
+ "loss": 0.2625,
211
+ "step": 575
212
+ },
213
+ {
214
+ "epoch": 0.48115477145148355,
215
+ "grad_norm": 0.17461912333965302,
216
+ "learning_rate": 0.0009414957031282751,
217
+ "loss": 0.2649,
218
+ "step": 600
219
+ },
220
+ {
221
+ "epoch": 0.48115477145148355,
222
+ "eval_loss": 0.24944312870502472,
223
+ "eval_runtime": 2.0116,
224
+ "eval_samples_per_second": 51.701,
225
+ "eval_steps_per_second": 0.994,
226
+ "step": 600
227
+ },
228
+ {
229
+ "epoch": 0.5012028869286287,
230
+ "grad_norm": 0.17447619140148163,
231
+ "learning_rate": 0.0009330127018922195,
232
+ "loss": 0.2597,
233
+ "step": 625
234
+ },
235
+ {
236
+ "epoch": 0.5212510024057738,
237
+ "grad_norm": 0.1970607340335846,
238
+ "learning_rate": 0.0009239993167235614,
239
+ "loss": 0.2577,
240
+ "step": 650
241
+ },
242
+ {
243
+ "epoch": 0.541299117882919,
244
+ "grad_norm": 0.17549267411231995,
245
+ "learning_rate": 0.0009144665878404079,
246
+ "loss": 0.2564,
247
+ "step": 675
248
+ },
249
+ {
250
+ "epoch": 0.5613472333600642,
251
+ "grad_norm": 0.17851398885250092,
252
+ "learning_rate": 0.0009044261915889984,
253
+ "loss": 0.2604,
254
+ "step": 700
255
+ },
256
+ {
257
+ "epoch": 0.5613472333600642,
258
+ "eval_loss": 0.24154677987098694,
259
+ "eval_runtime": 2.0084,
260
+ "eval_samples_per_second": 51.784,
261
+ "eval_steps_per_second": 0.996,
262
+ "step": 700
263
+ },
264
+ {
265
+ "epoch": 0.5813953488372093,
266
+ "grad_norm": 0.20294925570487976,
267
+ "learning_rate": 0.0008938904261417087,
268
+ "loss": 0.271,
269
+ "step": 725
270
+ },
271
+ {
272
+ "epoch": 0.6014434643143545,
273
+ "grad_norm": 0.16158199310302734,
274
+ "learning_rate": 0.0008828721964333975,
275
+ "loss": 0.2506,
276
+ "step": 750
277
+ },
278
+ {
279
+ "epoch": 0.6214915797914996,
280
+ "grad_norm": 0.2007351964712143,
281
+ "learning_rate": 0.000871384998354549,
282
+ "loss": 0.2454,
283
+ "step": 775
284
+ },
285
+ {
286
+ "epoch": 0.6415396952686447,
287
+ "grad_norm": 0.17736631631851196,
288
+ "learning_rate": 0.0008594429022205719,
289
+ "loss": 0.2514,
290
+ "step": 800
291
+ },
292
+ {
293
+ "epoch": 0.6415396952686447,
294
+ "eval_loss": 0.23755024373531342,
295
+ "eval_runtime": 4.2146,
296
+ "eval_samples_per_second": 24.676,
297
+ "eval_steps_per_second": 0.475,
298
+ "step": 800
299
+ },
300
+ {
301
+ "epoch": 0.6615878107457899,
302
+ "grad_norm": 0.15997523069381714,
303
+ "learning_rate": 0.0008470605355375032,
304
+ "loss": 0.2565,
305
+ "step": 825
306
+ },
307
+ {
308
+ "epoch": 0.681635926222935,
309
+ "grad_norm": 0.15349100530147552,
310
+ "learning_rate": 0.0008342530650852265,
311
+ "loss": 0.2598,
312
+ "step": 850
313
+ },
314
+ {
315
+ "epoch": 0.7016840417000801,
316
+ "grad_norm": 0.14758522808551788,
317
+ "learning_rate": 0.0008210361783401491,
318
+ "loss": 0.2426,
319
+ "step": 875
320
+ },
321
+ {
322
+ "epoch": 0.7217321571772254,
323
+ "grad_norm": 0.1686255782842636,
324
+ "learning_rate": 0.0008074260642600964,
325
+ "loss": 0.2462,
326
+ "step": 900
327
+ },
328
+ {
329
+ "epoch": 0.7217321571772254,
330
+ "eval_loss": 0.23332656919956207,
331
+ "eval_runtime": 4.2091,
332
+ "eval_samples_per_second": 24.708,
333
+ "eval_steps_per_second": 0.475,
334
+ "step": 900
335
+ },
336
+ {
337
+ "epoch": 0.7417802726543705,
338
+ "grad_norm": 0.1921042650938034,
339
+ "learning_rate": 0.0007934393934549542,
340
+ "loss": 0.249,
341
+ "step": 925
342
+ },
343
+ {
344
+ "epoch": 0.7618283881315157,
345
+ "grad_norm": 0.1523015797138214,
346
+ "learning_rate": 0.0007790932977673523,
347
+ "loss": 0.2399,
348
+ "step": 950
349
+ },
350
+ {
351
+ "epoch": 0.7818765036086608,
352
+ "grad_norm": 0.16055895388126373,
353
+ "learning_rate": 0.0007644053492883989,
354
+ "loss": 0.2363,
355
+ "step": 975
356
+ },
357
+ {
358
+ "epoch": 0.8019246190858059,
359
+ "grad_norm": 0.17422834038734436,
360
+ "learning_rate": 0.000749393538834164,
361
+ "loss": 0.2384,
362
+ "step": 1000
363
+ },
364
+ {
365
+ "epoch": 0.8019246190858059,
366
+ "eval_loss": 0.23113039135932922,
367
+ "eval_runtime": 4.2589,
368
+ "eval_samples_per_second": 24.419,
369
+ "eval_steps_per_second": 0.47,
370
+ "step": 1000
371
+ },
372
+ {
373
+ "epoch": 0.8219727345629511,
374
+ "grad_norm": 0.1599196046590805,
375
+ "learning_rate": 0.0007340762539092858,
376
+ "loss": 0.2485,
377
+ "step": 1025
378
+ },
379
+ {
380
+ "epoch": 0.8420208500400962,
381
+ "grad_norm": 0.15405167639255524,
382
+ "learning_rate": 0.0007184722561846798,
383
+ "loss": 0.2464,
384
+ "step": 1050
385
+ },
386
+ {
387
+ "epoch": 0.8620689655172413,
388
+ "grad_norm": 0.18259042501449585,
389
+ "learning_rate": 0.0007026006585169466,
390
+ "loss": 0.2421,
391
+ "step": 1075
392
+ },
393
+ {
394
+ "epoch": 0.8821170809943866,
395
+ "grad_norm": 0.14938652515411377,
396
+ "learning_rate": 0.0006864809015376217,
397
+ "loss": 0.2425,
398
+ "step": 1100
399
+ },
400
+ {
401
+ "epoch": 0.8821170809943866,
402
+ "eval_loss": 0.22659502923488617,
403
+ "eval_runtime": 4.2121,
404
+ "eval_samples_per_second": 24.691,
405
+ "eval_steps_per_second": 0.475,
406
+ "step": 1100
407
+ },
408
+ {
409
+ "epoch": 0.9021651964715317,
410
+ "grad_norm": 0.16447846591472626,
411
+ "learning_rate": 0.0006701327298409448,
412
+ "loss": 0.2414,
413
+ "step": 1125
414
+ },
415
+ {
416
+ "epoch": 0.9222133119486768,
417
+ "grad_norm": 0.1590721160173416,
418
+ "learning_rate": 0.000653576167799312,
419
+ "loss": 0.2287,
420
+ "step": 1150
421
+ },
422
+ {
423
+ "epoch": 0.942261427425822,
424
+ "grad_norm": 0.1653919219970703,
425
+ "learning_rate": 0.0006368314950360416,
426
+ "loss": 0.2351,
427
+ "step": 1175
428
+ },
429
+ {
430
+ "epoch": 0.9623095429029671,
431
+ "grad_norm": 0.1875888705253601,
432
+ "learning_rate": 0.000619919221585484,
433
+ "loss": 0.2374,
434
+ "step": 1200
435
+ },
436
+ {
437
+ "epoch": 0.9623095429029671,
438
+ "eval_loss": 0.22220070660114288,
439
+ "eval_runtime": 4.2412,
440
+ "eval_samples_per_second": 24.521,
441
+ "eval_steps_per_second": 0.472,
442
+ "step": 1200
443
+ },
444
+ {
445
+ "epoch": 0.9823576583801122,
446
+ "grad_norm": 0.15826693177223206,
447
+ "learning_rate": 0.0006028600627709151,
448
+ "loss": 0.2314,
449
+ "step": 1225
450
+ },
451
+ {
452
+ "epoch": 1.0024057738572574,
453
+ "grad_norm": 0.15599651634693146,
454
+ "learning_rate": 0.0005856749138309716,
455
+ "loss": 0.2246,
456
+ "step": 1250
457
+ },
458
+ {
459
+ "epoch": 1.0224538893344026,
460
+ "grad_norm": 0.15733949840068817,
461
+ "learning_rate": 0.000568384824325718,
462
+ "loss": 0.2024,
463
+ "step": 1275
464
+ },
465
+ {
466
+ "epoch": 1.0425020048115476,
467
+ "grad_norm": 0.14656169712543488,
468
+ "learning_rate": 0.0005510109723536876,
469
+ "loss": 0.2109,
470
+ "step": 1300
471
+ },
472
+ {
473
+ "epoch": 1.0425020048115476,
474
+ "eval_loss": 0.22098909318447113,
475
+ "eval_runtime": 4.2171,
476
+ "eval_samples_per_second": 24.661,
477
+ "eval_steps_per_second": 0.474,
478
+ "step": 1300
479
+ },
480
+ {
481
+ "epoch": 1.062550120288693,
482
+ "grad_norm": 0.15239104628562927,
483
+ "learning_rate": 0.0005335746386114814,
484
+ "loss": 0.1941,
485
+ "step": 1325
486
+ },
487
+ {
488
+ "epoch": 1.082598235765838,
489
+ "grad_norm": 0.1525331288576126,
490
+ "learning_rate": 0.0005160971803276981,
491
+ "loss": 0.2074,
492
+ "step": 1350
493
+ },
494
+ {
495
+ "epoch": 1.1026463512429832,
496
+ "grad_norm": 0.18646268546581268,
497
+ "learning_rate": 0.0004986000051031212,
498
+ "loss": 0.2008,
499
+ "step": 1375
500
+ },
501
+ {
502
+ "epoch": 1.1226944667201284,
503
+ "grad_norm": 0.1695125252008438,
504
+ "learning_rate": 0.00048110454468920866,
505
+ "loss": 0.2019,
506
+ "step": 1400
507
+ },
508
+ {
509
+ "epoch": 1.1226944667201284,
510
+ "eval_loss": 0.21638630330562592,
511
+ "eval_runtime": 4.2187,
512
+ "eval_samples_per_second": 24.652,
513
+ "eval_steps_per_second": 0.474,
514
+ "step": 1400
515
+ },
516
+ {
517
+ "epoch": 1.1427425821972734,
518
+ "grad_norm": 0.15290401875972748,
519
+ "learning_rate": 0.0004636322287369997,
520
+ "loss": 0.2021,
521
+ "step": 1425
522
+ },
523
+ {
524
+ "epoch": 1.1627906976744187,
525
+ "grad_norm": 0.16536127030849457,
526
+ "learning_rate": 0.0004462044585485944,
527
+ "loss": 0.1972,
528
+ "step": 1450
529
+ },
530
+ {
531
+ "epoch": 1.1828388131515637,
532
+ "grad_norm": 0.14967386424541473,
533
+ "learning_rate": 0.0004288425808633575,
534
+ "loss": 0.2033,
535
+ "step": 1475
536
+ },
537
+ {
538
+ "epoch": 1.202886928628709,
539
+ "grad_norm": 0.15446773171424866,
540
+ "learning_rate": 0.00041156786171095476,
541
+ "loss": 0.1957,
542
+ "step": 1500
543
+ },
544
+ {
545
+ "epoch": 1.202886928628709,
546
+ "eval_loss": 0.21511909365653992,
547
+ "eval_runtime": 4.2185,
548
+ "eval_samples_per_second": 24.654,
549
+ "eval_steps_per_second": 0.474,
550
+ "step": 1500
551
+ },
552
+ {
553
+ "epoch": 1.222935044105854,
554
+ "grad_norm": 0.1389647275209427,
555
+ "learning_rate": 0.00039440146036324753,
556
+ "loss": 0.1964,
557
+ "step": 1525
558
+ },
559
+ {
560
+ "epoch": 1.2429831595829992,
561
+ "grad_norm": 0.13631069660186768,
562
+ "learning_rate": 0.00037736440341695125,
563
+ "loss": 0.2033,
564
+ "step": 1550
565
+ },
566
+ {
567
+ "epoch": 1.2630312750601442,
568
+ "grad_norm": 0.1525258868932724,
569
+ "learning_rate": 0.0003604775590388047,
570
+ "loss": 0.2013,
571
+ "step": 1575
572
+ },
573
+ {
574
+ "epoch": 1.2830793905372895,
575
+ "grad_norm": 0.14098823070526123,
576
+ "learning_rate": 0.00034376161140479495,
577
+ "loss": 0.1976,
578
+ "step": 1600
579
+ },
580
+ {
581
+ "epoch": 1.2830793905372895,
582
+ "eval_loss": 0.21296119689941406,
583
+ "eval_runtime": 4.0201,
584
+ "eval_samples_per_second": 25.87,
585
+ "eval_steps_per_second": 0.497,
586
+ "step": 1600
587
+ },
588
+ {
589
+ "epoch": 1.3031275060144347,
590
+ "grad_norm": 0.1649613082408905,
591
+ "learning_rate": 0.0003272370353647465,
592
+ "loss": 0.2001,
593
+ "step": 1625
594
+ },
595
+ {
596
+ "epoch": 1.3231756214915797,
597
+ "grad_norm": 0.17128996551036835,
598
+ "learning_rate": 0.00031092407136330754,
599
+ "loss": 0.2015,
600
+ "step": 1650
601
+ },
602
+ {
603
+ "epoch": 1.343223736968725,
604
+ "grad_norm": 0.15200765430927277,
605
+ "learning_rate": 0.0002948427006480528,
606
+ "loss": 0.2056,
607
+ "step": 1675
608
+ },
609
+ {
610
+ "epoch": 1.36327185244587,
611
+ "grad_norm": 0.15264691412448883,
612
+ "learning_rate": 0.00027901262079506784,
613
+ "loss": 0.2032,
614
+ "step": 1700
615
+ },
616
+ {
617
+ "epoch": 1.36327185244587,
618
+ "eval_loss": 0.21197493374347687,
619
+ "eval_runtime": 4.256,
620
+ "eval_samples_per_second": 24.436,
621
+ "eval_steps_per_second": 0.47,
622
+ "step": 1700
623
+ },
624
+ {
625
+ "epoch": 1.3833199679230153,
626
+ "grad_norm": 0.16756217181682587,
627
+ "learning_rate": 0.000263453221581995,
628
+ "loss": 0.1987,
629
+ "step": 1725
630
+ },
631
+ {
632
+ "epoch": 1.4033680834001605,
633
+ "grad_norm": 0.13861249387264252,
634
+ "learning_rate": 0.00024818356123809036,
635
+ "loss": 0.1998,
636
+ "step": 1750
637
+ },
638
+ {
639
+ "epoch": 1.4234161988773055,
640
+ "grad_norm": 0.16581584513187408,
641
+ "learning_rate": 0.00023322234310038588,
642
+ "loss": 0.1875,
643
+ "step": 1775
644
+ },
645
+ {
646
+ "epoch": 1.4434643143544506,
647
+ "grad_norm": 0.1450669765472412,
648
+ "learning_rate": 0.00021858789270454783,
649
+ "loss": 0.1953,
650
+ "step": 1800
651
+ },
652
+ {
653
+ "epoch": 1.4434643143544506,
654
+ "eval_loss": 0.2067786008119583,
655
+ "eval_runtime": 4.2236,
656
+ "eval_samples_per_second": 24.624,
657
+ "eval_steps_per_second": 0.474,
658
+ "step": 1800
659
+ },
660
+ {
661
+ "epoch": 1.4635124298315958,
662
+ "grad_norm": 0.1343117356300354,
663
+ "learning_rate": 0.00020429813533849174,
664
+ "loss": 0.2026,
665
+ "step": 1825
666
+ },
667
+ {
668
+ "epoch": 1.483560545308741,
669
+ "grad_norm": 0.14949767291545868,
670
+ "learning_rate": 0.00019037057408624846,
671
+ "loss": 0.1929,
672
+ "step": 1850
673
+ },
674
+ {
675
+ "epoch": 1.5036086607858863,
676
+ "grad_norm": 0.12897883355617523,
677
+ "learning_rate": 0.00017682226838897568,
678
+ "loss": 0.1907,
679
+ "step": 1875
680
+ },
681
+ {
682
+ "epoch": 1.5236567762630313,
683
+ "grad_norm": 0.15060247480869293,
684
+ "learning_rate": 0.00016366981314937373,
685
+ "loss": 0.2003,
686
+ "step": 1900
687
+ },
688
+ {
689
+ "epoch": 1.5236567762630313,
690
+ "eval_loss": 0.2053409218788147,
691
+ "eval_runtime": 4.2538,
692
+ "eval_samples_per_second": 24.449,
693
+ "eval_steps_per_second": 0.47,
694
+ "step": 1900
695
+ },
696
+ {
697
+ "epoch": 1.5437048917401763,
698
+ "grad_norm": 0.1608167141675949,
699
+ "learning_rate": 0.0001509293184050995,
700
+ "loss": 0.1984,
701
+ "step": 1925
702
+ },
703
+ {
704
+ "epoch": 1.5637530072173216,
705
+ "grad_norm": 0.15426403284072876,
706
+ "learning_rate": 0.000138616389596077,
707
+ "loss": 0.1955,
708
+ "step": 1950
709
+ },
710
+ {
711
+ "epoch": 1.5838011226944668,
712
+ "grad_norm": 0.1431884467601776,
713
+ "learning_rate": 0.0001267461084498744,
714
+ "loss": 0.1955,
715
+ "step": 1975
716
+ },
717
+ {
718
+ "epoch": 1.6038492381716118,
719
+ "grad_norm": 0.14500346779823303,
720
+ "learning_rate": 0.00011533301450856055,
721
+ "loss": 0.1898,
722
+ "step": 2000
723
+ },
724
+ {
725
+ "epoch": 1.6038492381716118,
726
+ "eval_loss": 0.20465601980686188,
727
+ "eval_runtime": 4.2165,
728
+ "eval_samples_per_second": 24.665,
729
+ "eval_steps_per_second": 0.474,
730
+ "step": 2000
731
+ },
732
+ {
733
+ "epoch": 1.6238973536487569,
734
+ "grad_norm": 0.13658447563648224,
735
+ "learning_rate": 0.0001043910873196668,
736
+ "loss": 0.1882,
737
+ "step": 2025
738
+ },
739
+ {
740
+ "epoch": 1.6439454691259021,
741
+ "grad_norm": 0.14735296368598938,
742
+ "learning_rate": 9.393372931306943e-05,
743
+ "loss": 0.194,
744
+ "step": 2050
745
+ },
746
+ {
747
+ "epoch": 1.6639935846030474,
748
+ "grad_norm": 0.13270524144172668,
749
+ "learning_rate": 8.397374938476593e-05,
750
+ "loss": 0.191,
751
+ "step": 2075
752
+ },
753
+ {
754
+ "epoch": 1.6840417000801926,
755
+ "grad_norm": 0.13621263206005096,
756
+ "learning_rate": 7.452334720765258e-05,
757
+ "loss": 0.191,
758
+ "step": 2100
759
+ },
760
+ {
761
+ "epoch": 1.6840417000801926,
762
+ "eval_loss": 0.2029379904270172,
763
+ "eval_runtime": 4.2107,
764
+ "eval_samples_per_second": 24.699,
765
+ "eval_steps_per_second": 0.475,
766
+ "step": 2100
767
+ },
768
+ {
769
+ "epoch": 1.7040898155573376,
770
+ "grad_norm": 0.1418534368276596,
771
+ "learning_rate": 6.55940982885207e-05,
772
+ "loss": 0.1982,
773
+ "step": 2125
774
+ },
775
+ {
776
+ "epoch": 1.7241379310344827,
777
+ "grad_norm": 0.15418624877929688,
778
+ "learning_rate": 5.71969397895738e-05,
779
+ "loss": 0.1957,
780
+ "step": 2150
781
+ },
782
+ {
783
+ "epoch": 1.744186046511628,
784
+ "grad_norm": 0.15846756100654602,
785
+ "learning_rate": 4.934215713183526e-05,
786
+ "loss": 0.1904,
787
+ "step": 2175
788
+ },
789
+ {
790
+ "epoch": 1.7642341619887731,
791
+ "grad_norm": 0.15319091081619263,
792
+ "learning_rate": 4.203937139685188e-05,
793
+ "loss": 0.1835,
794
+ "step": 2200
795
+ },
796
+ {
797
+ "epoch": 1.7642341619887731,
798
+ "eval_loss": 0.20121867954730988,
799
+ "eval_runtime": 4.2209,
800
+ "eval_samples_per_second": 24.64,
801
+ "eval_steps_per_second": 0.474,
802
+ "step": 2200
803
+ },
804
+ {
805
+ "epoch": 1.7842822774659182,
806
+ "grad_norm": 0.14038674533367157,
807
+ "learning_rate": 3.529752754212767e-05,
808
+ "loss": 0.1886,
809
+ "step": 2225
810
+ },
811
+ {
812
+ "epoch": 1.8043303929430632,
813
+ "grad_norm": 0.13600395619869232,
814
+ "learning_rate": 2.9124883444720253e-05,
815
+ "loss": 0.186,
816
+ "step": 2250
817
+ },
818
+ {
819
+ "epoch": 1.8243785084202084,
820
+ "grad_norm": 0.1356409341096878,
821
+ "learning_rate": 2.3528999786421755e-05,
822
+ "loss": 0.183,
823
+ "step": 2275
824
+ },
825
+ {
826
+ "epoch": 1.8444266238973537,
827
+ "grad_norm": 0.14992156624794006,
828
+ "learning_rate": 1.851673079291216e-05,
829
+ "loss": 0.1738,
830
+ "step": 2300
831
+ },
832
+ {
833
+ "epoch": 1.8444266238973537,
834
+ "eval_loss": 0.20082467794418335,
835
+ "eval_runtime": 4.2146,
836
+ "eval_samples_per_second": 24.676,
837
+ "eval_steps_per_second": 0.475,
838
+ "step": 2300
839
+ },
840
+ {
841
+ "epoch": 1.864474739374499,
842
+ "grad_norm": 0.1463785320520401,
843
+ "learning_rate": 1.4094215838229174e-05,
844
+ "loss": 0.1982,
845
+ "step": 2325
846
+ },
847
+ {
848
+ "epoch": 1.884522854851644,
849
+ "grad_norm": 0.15394070744514465,
850
+ "learning_rate": 1.0266871924838216e-05,
851
+ "loss": 0.1887,
852
+ "step": 2350
853
+ },
854
+ {
855
+ "epoch": 1.904570970328789,
856
+ "grad_norm": 0.14152726531028748,
857
+ "learning_rate": 7.03938704851248e-06,
858
+ "loss": 0.1783,
859
+ "step": 2375
860
+ },
861
+ {
862
+ "epoch": 1.9246190858059342,
863
+ "grad_norm": 0.1539337933063507,
864
+ "learning_rate": 4.415714456151243e-06,
865
+ "loss": 0.182,
866
+ "step": 2400
867
+ },
868
+ {
869
+ "epoch": 1.9246190858059342,
870
+ "eval_loss": 0.20044730603694916,
871
+ "eval_runtime": 2.0109,
872
+ "eval_samples_per_second": 51.717,
873
+ "eval_steps_per_second": 0.995,
874
+ "step": 2400
875
+ },
876
+ {
877
+ "epoch": 1.9446672012830795,
878
+ "grad_norm": 0.13150149583816528,
879
+ "learning_rate": 2.3990678035694656e-06,
880
+ "loss": 0.1947,
881
+ "step": 2425
882
+ },
883
+ {
884
+ "epoch": 1.9647153167602245,
885
+ "grad_norm": 0.14985321462154388,
886
+ "learning_rate": 9.919172191896753e-07,
887
+ "loss": 0.1947,
888
+ "step": 2450
889
+ },
890
+ {
891
+ "epoch": 1.9847634322373697,
892
+ "grad_norm": 0.1531253457069397,
893
+ "learning_rate": 1.9598627845779372e-07,
894
+ "loss": 0.1918,
895
+ "step": 2475
896
+ },
897
+ {
898
+ "epoch": 2.0,
899
+ "step": 2494,
900
+ "total_flos": 1.6216560612723917e+18,
901
+ "train_loss": 0.23998703379198943,
902
+ "train_runtime": 3791.0396,
903
+ "train_samples_per_second": 21.048,
904
+ "train_steps_per_second": 0.658
905
+ }
906
+ ],
907
+ "logging_steps": 25,
908
+ "max_steps": 2494,
909
+ "num_input_tokens_seen": 0,
910
+ "num_train_epochs": 2,
911
+ "save_steps": 0,
912
+ "stateful_callbacks": {
913
+ "TrainerControl": {
914
+ "args": {
915
+ "should_epoch_stop": false,
916
+ "should_evaluate": false,
917
+ "should_log": false,
918
+ "should_save": true,
919
+ "should_training_stop": true
920
+ },
921
+ "attributes": {}
922
+ }
923
+ },
924
+ "total_flos": 1.6216560612723917e+18,
925
+ "train_batch_size": 32,
926
+ "trial_name": null,
927
+ "trial_params": null
928
+ }
nl_tasks/exp_init/run_ex02/ft/adapter_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "T": 1.0,
3
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "inference_mode": false,
6
+ "layers_to_transform": null,
7
+ "modules_to_save": null,
8
+ "num_rotations": 1,
9
+ "peft_type": "ROTATION",
10
+ "r": 16,
11
+ "revision": null,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "target_modules_to_skip": null,
17
+ "task_type": "CAUSAL_LM"
18
+ }
nl_tasks/exp_init/run_ex02/ft/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[PAD]": 32000
3
+ }
nl_tasks/exp_init/run_ex02/ft/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "</s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "</s>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
nl_tasks/exp_init/run_ex02/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
nl_tasks/exp_init/run_ex02/ft/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
nl_tasks/exp_init/run_ex02/ft/tokenizer_config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "[PAD]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ }
38
+ },
39
+ "bos_token": "</s>",
40
+ "clean_up_tokenization_spaces": false,
41
+ "eos_token": "</s>",
42
+ "extra_special_tokens": {},
43
+ "legacy": false,
44
+ "model_max_length": 512,
45
+ "pad_token": "[PAD]",
46
+ "padding_side": "right",
47
+ "sp_model_kwargs": {},
48
+ "tokenizer_class": "LlamaTokenizer",
49
+ "unk_token": "</s>",
50
+ "use_default_system_prompt": false
51
+ }
nl_tasks/exp_init/run_ex02/ft2/adapter_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "T": 1.0,
3
+ "base_model_name_or_path": "meta-llama/Llama-2-7b-hf",
4
+ "bias": "none",
5
+ "inference_mode": true,
6
+ "layers_to_transform": null,
7
+ "modules_to_save": null,
8
+ "num_rotations": 1,
9
+ "peft_type": "ROTATION",
10
+ "r": 16,
11
+ "revision": null,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "target_modules_to_skip": null,
17
+ "task_type": "CAUSAL_LM"
18
+ }
nl_tasks/exp_init/run_ex02/ft2/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c3f303f8fb6d50c04102ed0d1a7602f8dd61b95add6eccc0ce7ecd839ae4b25
3
+ size 33602915
nl_tasks/exp_init/run_ex02/trainer_state.json ADDED
@@ -0,0 +1,914 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.0,
6
+ "eval_steps": 100,
7
+ "global_step": 2438,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.020508613617719443,
14
+ "grad_norm": 0.3222310543060303,
15
+ "learning_rate": 9.836065573770491e-05,
16
+ "loss": 0.6045,
17
+ "step": 25
18
+ },
19
+ {
20
+ "epoch": 0.04101722723543889,
21
+ "grad_norm": 0.3179238736629486,
22
+ "learning_rate": 0.00020081967213114754,
23
+ "loss": 0.3855,
24
+ "step": 50
25
+ },
26
+ {
27
+ "epoch": 0.06152584085315833,
28
+ "grad_norm": 0.29217585921287537,
29
+ "learning_rate": 0.0003032786885245902,
30
+ "loss": 0.3566,
31
+ "step": 75
32
+ },
33
+ {
34
+ "epoch": 0.08203445447087777,
35
+ "grad_norm": 0.3154868185520172,
36
+ "learning_rate": 0.0004057377049180328,
37
+ "loss": 0.3237,
38
+ "step": 100
39
+ },
40
+ {
41
+ "epoch": 0.08203445447087777,
42
+ "eval_loss": 0.31438109278678894,
43
+ "eval_runtime": 19.4649,
44
+ "eval_samples_per_second": 51.375,
45
+ "eval_steps_per_second": 0.822,
46
+ "step": 100
47
+ },
48
+ {
49
+ "epoch": 0.10254306808859721,
50
+ "grad_norm": 0.3041672706604004,
51
+ "learning_rate": 0.0005081967213114754,
52
+ "loss": 0.3306,
53
+ "step": 125
54
+ },
55
+ {
56
+ "epoch": 0.12305168170631665,
57
+ "grad_norm": 0.3167073428630829,
58
+ "learning_rate": 0.000610655737704918,
59
+ "loss": 0.3149,
60
+ "step": 150
61
+ },
62
+ {
63
+ "epoch": 0.1435602953240361,
64
+ "grad_norm": 0.3177391290664673,
65
+ "learning_rate": 0.0007131147540983607,
66
+ "loss": 0.3063,
67
+ "step": 175
68
+ },
69
+ {
70
+ "epoch": 0.16406890894175555,
71
+ "grad_norm": 0.33738526701927185,
72
+ "learning_rate": 0.0008155737704918033,
73
+ "loss": 0.3202,
74
+ "step": 200
75
+ },
76
+ {
77
+ "epoch": 0.16406890894175555,
78
+ "eval_loss": 0.30314716696739197,
79
+ "eval_runtime": 19.2923,
80
+ "eval_samples_per_second": 51.834,
81
+ "eval_steps_per_second": 0.829,
82
+ "step": 200
83
+ },
84
+ {
85
+ "epoch": 0.184577522559475,
86
+ "grad_norm": 0.30629661679267883,
87
+ "learning_rate": 0.0009180327868852459,
88
+ "loss": 0.3179,
89
+ "step": 225
90
+ },
91
+ {
92
+ "epoch": 0.20508613617719443,
93
+ "grad_norm": 0.9242148995399475,
94
+ "learning_rate": 0.0009999871854116063,
95
+ "loss": 0.2964,
96
+ "step": 250
97
+ },
98
+ {
99
+ "epoch": 0.22559474979491387,
100
+ "grad_norm": 0.2491447776556015,
101
+ "learning_rate": 0.0009995387437838027,
102
+ "loss": 0.2987,
103
+ "step": 275
104
+ },
105
+ {
106
+ "epoch": 0.2461033634126333,
107
+ "grad_norm": 0.33308619260787964,
108
+ "learning_rate": 0.000998450229439693,
109
+ "loss": 0.2963,
110
+ "step": 300
111
+ },
112
+ {
113
+ "epoch": 0.2461033634126333,
114
+ "eval_loss": 0.2860982120037079,
115
+ "eval_runtime": 19.2995,
116
+ "eval_samples_per_second": 51.815,
117
+ "eval_steps_per_second": 0.829,
118
+ "step": 300
119
+ },
120
+ {
121
+ "epoch": 0.2666119770303528,
122
+ "grad_norm": 0.2559486925601959,
123
+ "learning_rate": 0.000996723037122612,
124
+ "loss": 0.2973,
125
+ "step": 325
126
+ },
127
+ {
128
+ "epoch": 0.2871205906480722,
129
+ "grad_norm": 0.219248428940773,
130
+ "learning_rate": 0.0009943593799315263,
131
+ "loss": 0.2889,
132
+ "step": 350
133
+ },
134
+ {
135
+ "epoch": 0.30762920426579166,
136
+ "grad_norm": 0.2518327236175537,
137
+ "learning_rate": 0.0009913622864853324,
138
+ "loss": 0.2972,
139
+ "step": 375
140
+ },
141
+ {
142
+ "epoch": 0.3281378178835111,
143
+ "grad_norm": 0.2639095187187195,
144
+ "learning_rate": 0.0009877355970422024,
145
+ "loss": 0.2823,
146
+ "step": 400
147
+ },
148
+ {
149
+ "epoch": 0.3281378178835111,
150
+ "eval_loss": 0.27193590998649597,
151
+ "eval_runtime": 19.297,
152
+ "eval_samples_per_second": 51.822,
153
+ "eval_steps_per_second": 0.829,
154
+ "step": 400
155
+ },
156
+ {
157
+ "epoch": 0.34864643150123054,
158
+ "grad_norm": 0.21691828966140747,
159
+ "learning_rate": 0.0009834839585789559,
160
+ "loss": 0.2888,
161
+ "step": 425
162
+ },
163
+ {
164
+ "epoch": 0.36915504511895,
165
+ "grad_norm": 0.2310950607061386,
166
+ "learning_rate": 0.000978612818836762,
167
+ "loss": 0.2844,
168
+ "step": 450
169
+ },
170
+ {
171
+ "epoch": 0.3896636587366694,
172
+ "grad_norm": 0.21057192981243134,
173
+ "learning_rate": 0.0009731284193407981,
174
+ "loss": 0.2733,
175
+ "step": 475
176
+ },
177
+ {
178
+ "epoch": 0.41017227235438886,
179
+ "grad_norm": 0.22767874598503113,
180
+ "learning_rate": 0.0009670377874028117,
181
+ "loss": 0.2574,
182
+ "step": 500
183
+ },
184
+ {
185
+ "epoch": 0.41017227235438886,
186
+ "eval_loss": 0.26140961050987244,
187
+ "eval_runtime": 19.307,
188
+ "eval_samples_per_second": 51.795,
189
+ "eval_steps_per_second": 0.829,
190
+ "step": 500
191
+ },
192
+ {
193
+ "epoch": 0.4306808859721083,
194
+ "grad_norm": 0.20047912001609802,
195
+ "learning_rate": 0.0009603487271168336,
196
+ "loss": 0.2593,
197
+ "step": 525
198
+ },
199
+ {
200
+ "epoch": 0.45118949958982774,
201
+ "grad_norm": 0.19039607048034668,
202
+ "learning_rate": 0.0009530698093595781,
203
+ "loss": 0.2696,
204
+ "step": 550
205
+ },
206
+ {
207
+ "epoch": 0.4716981132075472,
208
+ "grad_norm": 0.20591352880001068,
209
+ "learning_rate": 0.0009452103608083418,
210
+ "loss": 0.2604,
211
+ "step": 575
212
+ },
213
+ {
214
+ "epoch": 0.4922067268252666,
215
+ "grad_norm": 0.2514936625957489,
216
+ "learning_rate": 0.0009367804519904775,
217
+ "loss": 0.2784,
218
+ "step": 600
219
+ },
220
+ {
221
+ "epoch": 0.4922067268252666,
222
+ "eval_loss": 0.2537764608860016,
223
+ "eval_runtime": 19.2927,
224
+ "eval_samples_per_second": 51.833,
225
+ "eval_steps_per_second": 0.829,
226
+ "step": 600
227
+ },
228
+ {
229
+ "epoch": 0.5127153404429861,
230
+ "grad_norm": 0.1930495798587799,
231
+ "learning_rate": 0.0009277908843797492,
232
+ "loss": 0.2562,
233
+ "step": 625
234
+ },
235
+ {
236
+ "epoch": 0.5332239540607056,
237
+ "grad_norm": 0.18388758599758148,
238
+ "learning_rate": 0.0009182531765561084,
239
+ "loss": 0.2503,
240
+ "step": 650
241
+ },
242
+ {
243
+ "epoch": 0.5537325676784249,
244
+ "grad_norm": 0.17093320190906525,
245
+ "learning_rate": 0.0009081795494466201,
246
+ "loss": 0.2632,
247
+ "step": 675
248
+ },
249
+ {
250
+ "epoch": 0.5742411812961444,
251
+ "grad_norm": 0.171810120344162,
252
+ "learning_rate": 0.0008975829106664539,
253
+ "loss": 0.2605,
254
+ "step": 700
255
+ },
256
+ {
257
+ "epoch": 0.5742411812961444,
258
+ "eval_loss": 0.2481871396303177,
259
+ "eval_runtime": 19.293,
260
+ "eval_samples_per_second": 51.832,
261
+ "eval_steps_per_second": 0.829,
262
+ "step": 700
263
+ },
264
+ {
265
+ "epoch": 0.5947497949138638,
266
+ "grad_norm": 0.16749516129493713,
267
+ "learning_rate": 0.0008864768379800017,
268
+ "loss": 0.2588,
269
+ "step": 725
270
+ },
271
+ {
272
+ "epoch": 0.6152584085315833,
273
+ "grad_norm": 0.17328386008739471,
274
+ "learning_rate": 0.0008748755619033153,
275
+ "loss": 0.2551,
276
+ "step": 750
277
+ },
278
+ {
279
+ "epoch": 0.6357670221493027,
280
+ "grad_norm": 0.17557500302791595,
281
+ "learning_rate": 0.000862793947470155,
282
+ "loss": 0.2527,
283
+ "step": 775
284
+ },
285
+ {
286
+ "epoch": 0.6562756357670222,
287
+ "grad_norm": 0.17228572070598602,
288
+ "learning_rate": 0.0008502474751850142,
289
+ "loss": 0.2546,
290
+ "step": 800
291
+ },
292
+ {
293
+ "epoch": 0.6562756357670222,
294
+ "eval_loss": 0.2438872903585434,
295
+ "eval_runtime": 19.3466,
296
+ "eval_samples_per_second": 51.689,
297
+ "eval_steps_per_second": 0.827,
298
+ "step": 800
299
+ },
300
+ {
301
+ "epoch": 0.6767842493847416,
302
+ "grad_norm": 0.18187439441680908,
303
+ "learning_rate": 0.0008372522211875224,
304
+ "loss": 0.2509,
305
+ "step": 825
306
+ },
307
+ {
308
+ "epoch": 0.6972928630024611,
309
+ "grad_norm": 0.15625031292438507,
310
+ "learning_rate": 0.0008238248366536473,
311
+ "loss": 0.2524,
312
+ "step": 850
313
+ },
314
+ {
315
+ "epoch": 0.7178014766201805,
316
+ "grad_norm": 0.14813965559005737,
317
+ "learning_rate": 0.0008099825264600842,
318
+ "loss": 0.2417,
319
+ "step": 875
320
+ },
321
+ {
322
+ "epoch": 0.7383100902379,
323
+ "grad_norm": 0.17053155601024628,
324
+ "learning_rate": 0.0007957430271391761,
325
+ "loss": 0.2478,
326
+ "step": 900
327
+ },
328
+ {
329
+ "epoch": 0.7383100902379,
330
+ "eval_loss": 0.23720978200435638,
331
+ "eval_runtime": 19.2795,
332
+ "eval_samples_per_second": 51.869,
333
+ "eval_steps_per_second": 0.83,
334
+ "step": 900
335
+ },
336
+ {
337
+ "epoch": 0.7588187038556193,
338
+ "grad_norm": 0.1528586894273758,
339
+ "learning_rate": 0.0007811245841526062,
340
+ "loss": 0.2491,
341
+ "step": 925
342
+ },
343
+ {
344
+ "epoch": 0.7793273174733388,
345
+ "grad_norm": 0.1788587123155594,
346
+ "learning_rate": 0.0007661459285129879,
347
+ "loss": 0.2373,
348
+ "step": 950
349
+ },
350
+ {
351
+ "epoch": 0.7998359310910582,
352
+ "grad_norm": 0.17576590180397034,
353
+ "learning_rate": 0.0007508262527833029,
354
+ "loss": 0.2446,
355
+ "step": 975
356
+ },
357
+ {
358
+ "epoch": 0.8203445447087777,
359
+ "grad_norm": 0.16797195374965668,
360
+ "learning_rate": 0.000735185186484943,
361
+ "loss": 0.2377,
362
+ "step": 1000
363
+ },
364
+ {
365
+ "epoch": 0.8203445447087777,
366
+ "eval_loss": 0.23422712087631226,
367
+ "eval_runtime": 19.2981,
368
+ "eval_samples_per_second": 51.819,
369
+ "eval_steps_per_second": 0.829,
370
+ "step": 1000
371
+ },
372
+ {
373
+ "epoch": 0.8408531583264971,
374
+ "grad_norm": 0.16894708573818207,
375
+ "learning_rate": 0.0007192427709458656,
376
+ "loss": 0.2378,
377
+ "step": 1025
378
+ },
379
+ {
380
+ "epoch": 0.8613617719442166,
381
+ "grad_norm": 0.15344825387001038,
382
+ "learning_rate": 0.0007030194336210887,
383
+ "loss": 0.2362,
384
+ "step": 1050
385
+ },
386
+ {
387
+ "epoch": 0.881870385561936,
388
+ "grad_norm": 0.1591627299785614,
389
+ "learning_rate": 0.0006865359619184331,
390
+ "loss": 0.2407,
391
+ "step": 1075
392
+ },
393
+ {
394
+ "epoch": 0.9023789991796555,
395
+ "grad_norm": 0.1595323532819748,
396
+ "learning_rate": 0.0006698134765630434,
397
+ "loss": 0.2401,
398
+ "step": 1100
399
+ },
400
+ {
401
+ "epoch": 0.9023789991796555,
402
+ "eval_loss": 0.2298676073551178,
403
+ "eval_runtime": 19.2888,
404
+ "eval_samples_per_second": 51.844,
405
+ "eval_steps_per_second": 0.829,
406
+ "step": 1100
407
+ },
408
+ {
409
+ "epoch": 0.9228876127973749,
410
+ "grad_norm": 0.1466217339038849,
411
+ "learning_rate": 0.0006528734045348248,
412
+ "loss": 0.236,
413
+ "step": 1125
414
+ },
415
+ {
416
+ "epoch": 0.9433962264150944,
417
+ "grad_norm": 0.1441761553287506,
418
+ "learning_rate": 0.0006357374516134643,
419
+ "loss": 0.2346,
420
+ "step": 1150
421
+ },
422
+ {
423
+ "epoch": 0.9639048400328137,
424
+ "grad_norm": 0.13931916654109955,
425
+ "learning_rate": 0.0006184275745662179,
426
+ "loss": 0.2412,
427
+ "step": 1175
428
+ },
429
+ {
430
+ "epoch": 0.9844134536505332,
431
+ "grad_norm": 0.1619672328233719,
432
+ "learning_rate": 0.0006009659530141031,
433
+ "loss": 0.2458,
434
+ "step": 1200
435
+ },
436
+ {
437
+ "epoch": 0.9844134536505332,
438
+ "eval_loss": 0.22785130143165588,
439
+ "eval_runtime": 19.2959,
440
+ "eval_samples_per_second": 51.824,
441
+ "eval_steps_per_second": 0.829,
442
+ "step": 1200
443
+ },
444
+ {
445
+ "epoch": 1.0049220672682526,
446
+ "grad_norm": 0.1470383256673813,
447
+ "learning_rate": 0.0005833749610125402,
448
+ "loss": 0.2273,
449
+ "step": 1225
450
+ },
451
+ {
452
+ "epoch": 1.0254306808859721,
453
+ "grad_norm": 0.14438124001026154,
454
+ "learning_rate": 0.0005656771383828602,
455
+ "loss": 0.2068,
456
+ "step": 1250
457
+ },
458
+ {
459
+ "epoch": 1.0459392945036916,
460
+ "grad_norm": 0.16950909793376923,
461
+ "learning_rate": 0.0005478951618314134,
462
+ "loss": 0.207,
463
+ "step": 1275
464
+ },
465
+ {
466
+ "epoch": 1.066447908121411,
467
+ "grad_norm": 0.14760279655456543,
468
+ "learning_rate": 0.0005300518158932815,
469
+ "loss": 0.2096,
470
+ "step": 1300
471
+ },
472
+ {
473
+ "epoch": 1.066447908121411,
474
+ "eval_loss": 0.22524121403694153,
475
+ "eval_runtime": 19.2901,
476
+ "eval_samples_per_second": 51.84,
477
+ "eval_steps_per_second": 0.829,
478
+ "step": 1300
479
+ },
480
+ {
481
+ "epoch": 1.0869565217391304,
482
+ "grad_norm": 0.15871679782867432,
483
+ "learning_rate": 0.0005121699637378282,
484
+ "loss": 0.2074,
485
+ "step": 1325
486
+ },
487
+ {
488
+ "epoch": 1.1074651353568499,
489
+ "grad_norm": 0.14514082670211792,
490
+ "learning_rate": 0.0004942725178734903,
491
+ "loss": 0.2003,
492
+ "step": 1350
493
+ },
494
+ {
495
+ "epoch": 1.1279737489745694,
496
+ "grad_norm": 0.18950478732585907,
497
+ "learning_rate": 0.00047638241078935324,
498
+ "loss": 0.2024,
499
+ "step": 1375
500
+ },
501
+ {
502
+ "epoch": 1.1484823625922886,
503
+ "grad_norm": 0.1584726721048355,
504
+ "learning_rate": 0.000458522565571121,
505
+ "loss": 0.1955,
506
+ "step": 1400
507
+ },
508
+ {
509
+ "epoch": 1.1484823625922886,
510
+ "eval_loss": 0.22360759973526,
511
+ "eval_runtime": 19.3066,
512
+ "eval_samples_per_second": 51.796,
513
+ "eval_steps_per_second": 0.829,
514
+ "step": 1400
515
+ },
516
+ {
517
+ "epoch": 1.1689909762100081,
518
+ "grad_norm": 0.15903767943382263,
519
+ "learning_rate": 0.00044071586652913767,
520
+ "loss": 0.2069,
521
+ "step": 1425
522
+ },
523
+ {
524
+ "epoch": 1.1894995898277276,
525
+ "grad_norm": 0.15384770929813385,
526
+ "learning_rate": 0.0004229851298760915,
527
+ "loss": 0.206,
528
+ "step": 1450
529
+ },
530
+ {
531
+ "epoch": 1.2100082034454471,
532
+ "grad_norm": 0.16331611573696136,
533
+ "learning_rate": 0.0004053530744919749,
534
+ "loss": 0.2041,
535
+ "step": 1475
536
+ },
537
+ {
538
+ "epoch": 1.2305168170631666,
539
+ "grad_norm": 0.17139384150505066,
540
+ "learning_rate": 0.0003878422928137597,
541
+ "loss": 0.2127,
542
+ "step": 1500
543
+ },
544
+ {
545
+ "epoch": 1.2305168170631666,
546
+ "eval_loss": 0.2206723392009735,
547
+ "eval_runtime": 19.2698,
548
+ "eval_samples_per_second": 51.895,
549
+ "eval_steps_per_second": 0.83,
550
+ "step": 1500
551
+ },
552
+ {
553
+ "epoch": 1.251025430680886,
554
+ "grad_norm": 0.19464436173439026,
555
+ "learning_rate": 0.0003704752218870861,
556
+ "loss": 0.207,
557
+ "step": 1525
558
+ },
559
+ {
560
+ "epoch": 1.2715340442986054,
561
+ "grad_norm": 0.1387706995010376,
562
+ "learning_rate": 0.00035327411461706025,
563
+ "loss": 0.1999,
564
+ "step": 1550
565
+ },
566
+ {
567
+ "epoch": 1.2920426579163249,
568
+ "grad_norm": 0.137301504611969,
569
+ "learning_rate": 0.00033626101125499555,
570
+ "loss": 0.2026,
571
+ "step": 1575
572
+ },
573
+ {
574
+ "epoch": 1.3125512715340442,
575
+ "grad_norm": 0.1531214863061905,
576
+ "learning_rate": 0.0003194577111576333,
577
+ "loss": 0.1957,
578
+ "step": 1600
579
+ },
580
+ {
581
+ "epoch": 1.3125512715340442,
582
+ "eval_loss": 0.2175484299659729,
583
+ "eval_runtime": 19.2699,
584
+ "eval_samples_per_second": 51.894,
585
+ "eval_steps_per_second": 0.83,
586
+ "step": 1600
587
+ },
588
+ {
589
+ "epoch": 1.3330598851517639,
590
+ "grad_norm": 0.14256389439105988,
591
+ "learning_rate": 0.00030288574485502756,
592
+ "loss": 0.198,
593
+ "step": 1625
594
+ },
595
+ {
596
+ "epoch": 1.3535684987694832,
597
+ "grad_norm": 0.15813028812408447,
598
+ "learning_rate": 0.00028656634646288565,
599
+ "loss": 0.199,
600
+ "step": 1650
601
+ },
602
+ {
603
+ "epoch": 1.3740771123872026,
604
+ "grad_norm": 0.17235884070396423,
605
+ "learning_rate": 0.00027052042647471254,
606
+ "loss": 0.1959,
607
+ "step": 1675
608
+ },
609
+ {
610
+ "epoch": 1.3945857260049221,
611
+ "grad_norm": 0.14712415635585785,
612
+ "learning_rate": 0.0002547685449686206,
613
+ "loss": 0.1971,
614
+ "step": 1700
615
+ },
616
+ {
617
+ "epoch": 1.3945857260049221,
618
+ "eval_loss": 0.21476209163665771,
619
+ "eval_runtime": 19.2701,
620
+ "eval_samples_per_second": 51.894,
621
+ "eval_steps_per_second": 0.83,
622
+ "step": 1700
623
+ },
624
+ {
625
+ "epoch": 1.4150943396226414,
626
+ "grad_norm": 0.14806444942951202,
627
+ "learning_rate": 0.0002393308852631373,
628
+ "loss": 0.1936,
629
+ "step": 1725
630
+ },
631
+ {
632
+ "epoch": 1.435602953240361,
633
+ "grad_norm": 0.1419476568698883,
634
+ "learning_rate": 0.0002242272280557645,
635
+ "loss": 0.1929,
636
+ "step": 1750
637
+ },
638
+ {
639
+ "epoch": 1.4561115668580804,
640
+ "grad_norm": 0.14897122979164124,
641
+ "learning_rate": 0.0002094769260774262,
642
+ "loss": 0.1998,
643
+ "step": 1775
644
+ },
645
+ {
646
+ "epoch": 1.4766201804758,
647
+ "grad_norm": 0.13588683307170868,
648
+ "learning_rate": 0.00019509887929528458,
649
+ "loss": 0.1891,
650
+ "step": 1800
651
+ },
652
+ {
653
+ "epoch": 1.4766201804758,
654
+ "eval_loss": 0.21355417370796204,
655
+ "eval_runtime": 19.2809,
656
+ "eval_samples_per_second": 51.865,
657
+ "eval_steps_per_second": 0.83,
658
+ "step": 1800
659
+ },
660
+ {
661
+ "epoch": 1.4971287940935194,
662
+ "grad_norm": 0.13949081301689148,
663
+ "learning_rate": 0.0001811115106956918,
664
+ "loss": 0.1886,
665
+ "step": 1825
666
+ },
667
+ {
668
+ "epoch": 1.5176374077112387,
669
+ "grad_norm": 0.1364123523235321,
670
+ "learning_rate": 0.00016753274267831115,
671
+ "loss": 0.1916,
672
+ "step": 1850
673
+ },
674
+ {
675
+ "epoch": 1.5381460213289582,
676
+ "grad_norm": 0.15067338943481445,
677
+ "learning_rate": 0.00015437997409165478,
678
+ "loss": 0.1918,
679
+ "step": 1875
680
+ },
681
+ {
682
+ "epoch": 1.5586546349466777,
683
+ "grad_norm": 0.1649065762758255,
684
+ "learning_rate": 0.00014167005793946035,
685
+ "loss": 0.1956,
686
+ "step": 1900
687
+ },
688
+ {
689
+ "epoch": 1.5586546349466777,
690
+ "eval_loss": 0.21143756806850433,
691
+ "eval_runtime": 19.2693,
692
+ "eval_samples_per_second": 51.896,
693
+ "eval_steps_per_second": 0.83,
694
+ "step": 1900
695
+ },
696
+ {
697
+ "epoch": 1.579163248564397,
698
+ "grad_norm": 0.1650763899087906,
699
+ "learning_rate": 0.00012941927978647527,
700
+ "loss": 0.1965,
701
+ "step": 1925
702
+ },
703
+ {
704
+ "epoch": 1.5996718621821167,
705
+ "grad_norm": 0.1541653871536255,
706
+ "learning_rate": 0.00011764333689131385,
707
+ "loss": 0.194,
708
+ "step": 1950
709
+ },
710
+ {
711
+ "epoch": 1.620180475799836,
712
+ "grad_norm": 0.14436492323875427,
713
+ "learning_rate": 0.00010635731809312993,
714
+ "loss": 0.1906,
715
+ "step": 1975
716
+ },
717
+ {
718
+ "epoch": 1.6406890894175554,
719
+ "grad_norm": 0.15361854434013367,
720
+ "learning_rate": 9.557568447787201e-05,
721
+ "loss": 0.1958,
722
+ "step": 2000
723
+ },
724
+ {
725
+ "epoch": 1.6406890894175554,
726
+ "eval_loss": 0.2098333090543747,
727
+ "eval_runtime": 19.3433,
728
+ "eval_samples_per_second": 51.697,
729
+ "eval_steps_per_second": 0.827,
730
+ "step": 2000
731
+ },
732
+ {
733
+ "epoch": 1.661197703035275,
734
+ "grad_norm": 0.1444142609834671,
735
+ "learning_rate": 8.531225084889654e-05,
736
+ "loss": 0.1866,
737
+ "step": 2025
738
+ },
739
+ {
740
+ "epoch": 1.6817063166529942,
741
+ "grad_norm": 0.16099859774112701,
742
+ "learning_rate": 7.558016802568091e-05,
743
+ "loss": 0.1881,
744
+ "step": 2050
745
+ },
746
+ {
747
+ "epoch": 1.7022149302707137,
748
+ "grad_norm": 0.18300087749958038,
749
+ "learning_rate": 6.639190599331746e-05,
750
+ "loss": 0.1928,
751
+ "step": 2075
752
+ },
753
+ {
754
+ "epoch": 1.7227235438884332,
755
+ "grad_norm": 0.16143175959587097,
756
+ "learning_rate": 5.775923792437865e-05,
757
+ "loss": 0.1836,
758
+ "step": 2100
759
+ },
760
+ {
761
+ "epoch": 1.7227235438884332,
762
+ "eval_loss": 0.2088012397289276,
763
+ "eval_runtime": 19.3492,
764
+ "eval_samples_per_second": 51.682,
765
+ "eval_steps_per_second": 0.827,
766
+ "step": 2100
767
+ },
768
+ {
769
+ "epoch": 1.7432321575061525,
770
+ "grad_norm": 0.1365450620651245,
771
+ "learning_rate": 4.9693225093627616e-05,
772
+ "loss": 0.188,
773
+ "step": 2125
774
+ },
775
+ {
776
+ "epoch": 1.7637407711238722,
777
+ "grad_norm": 0.1199432909488678,
778
+ "learning_rate": 4.220420270490294e-05,
779
+ "loss": 0.1861,
780
+ "step": 2150
781
+ },
782
+ {
783
+ "epoch": 1.7842493847415914,
784
+ "grad_norm": 0.13452377915382385,
785
+ "learning_rate": 3.530176664833834e-05,
786
+ "loss": 0.1955,
787
+ "step": 2175
788
+ },
789
+ {
790
+ "epoch": 1.804757998359311,
791
+ "grad_norm": 0.15721169114112854,
792
+ "learning_rate": 2.8994761204884756e-05,
793
+ "loss": 0.188,
794
+ "step": 2200
795
+ },
796
+ {
797
+ "epoch": 1.804757998359311,
798
+ "eval_loss": 0.20743653178215027,
799
+ "eval_runtime": 19.302,
800
+ "eval_samples_per_second": 51.808,
801
+ "eval_steps_per_second": 0.829,
802
+ "step": 2200
803
+ },
804
+ {
805
+ "epoch": 1.8252666119770304,
806
+ "grad_norm": 0.15070787072181702,
807
+ "learning_rate": 2.329126771388995e-05,
808
+ "loss": 0.189,
809
+ "step": 2225
810
+ },
811
+ {
812
+ "epoch": 1.8457752255947497,
813
+ "grad_norm": 0.1434791386127472,
814
+ "learning_rate": 1.8198594218256815e-05,
815
+ "loss": 0.19,
816
+ "step": 2250
817
+ },
818
+ {
819
+ "epoch": 1.8662838392124692,
820
+ "grad_norm": 0.13482141494750977,
821
+ "learning_rate": 1.3723266100447052e-05,
822
+ "loss": 0.1962,
823
+ "step": 2275
824
+ },
825
+ {
826
+ "epoch": 1.8867924528301887,
827
+ "grad_norm": 0.13452687859535217,
828
+ "learning_rate": 9.871017721329201e-06,
829
+ "loss": 0.1881,
830
+ "step": 2300
831
+ },
832
+ {
833
+ "epoch": 1.8867924528301887,
834
+ "eval_loss": 0.20719286799430847,
835
+ "eval_runtime": 19.348,
836
+ "eval_samples_per_second": 51.685,
837
+ "eval_steps_per_second": 0.827,
838
+ "step": 2300
839
+ },
840
+ {
841
+ "epoch": 1.907301066447908,
842
+ "grad_norm": 0.16005028784275055,
843
+ "learning_rate": 6.646785072584871e-06,
844
+ "loss": 0.1958,
845
+ "step": 2325
846
+ },
847
+ {
848
+ "epoch": 1.9278096800656277,
849
+ "grad_norm": 0.1455722600221634,
850
+ "learning_rate": 4.054699452086641e-06,
851
+ "loss": 0.1816,
852
+ "step": 2350
853
+ },
854
+ {
855
+ "epoch": 1.948318293683347,
856
+ "grad_norm": 0.12664249539375305,
857
+ "learning_rate": 2.0980821703527886e-06,
858
+ "loss": 0.1867,
859
+ "step": 2375
860
+ },
861
+ {
862
+ "epoch": 1.9688269073010665,
863
+ "grad_norm": 0.14652538299560547,
864
+ "learning_rate": 7.794402948607671e-07,
865
+ "loss": 0.1876,
866
+ "step": 2400
867
+ },
868
+ {
869
+ "epoch": 1.9688269073010665,
870
+ "eval_loss": 0.20708177983760834,
871
+ "eval_runtime": 19.3185,
872
+ "eval_samples_per_second": 51.764,
873
+ "eval_steps_per_second": 0.828,
874
+ "step": 2400
875
+ },
876
+ {
877
+ "epoch": 1.989335520918786,
878
+ "grad_norm": 0.14083988964557648,
879
+ "learning_rate": 1.0046343767294853e-07,
880
+ "loss": 0.184,
881
+ "step": 2425
882
+ },
883
+ {
884
+ "epoch": 2.0,
885
+ "step": 2438,
886
+ "total_flos": 1.58523627405312e+18,
887
+ "train_loss": 0.23704243547315965,
888
+ "train_runtime": 2703.034,
889
+ "train_samples_per_second": 28.856,
890
+ "train_steps_per_second": 0.902
891
+ }
892
+ ],
893
+ "logging_steps": 25,
894
+ "max_steps": 2438,
895
+ "num_input_tokens_seen": 0,
896
+ "num_train_epochs": 2,
897
+ "save_steps": 0,
898
+ "stateful_callbacks": {
899
+ "TrainerControl": {
900
+ "args": {
901
+ "should_epoch_stop": false,
902
+ "should_evaluate": false,
903
+ "should_log": false,
904
+ "should_save": true,
905
+ "should_training_stop": true
906
+ },
907
+ "attributes": {}
908
+ }
909
+ },
910
+ "total_flos": 1.58523627405312e+18,
911
+ "train_batch_size": 32,
912
+ "trial_name": null,
913
+ "trial_params": null
914
+ }