|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 500.0, |
|
"eval_steps": 500, |
|
"global_step": 1000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 8.514540942565027, |
|
"learning_rate": 5e-06, |
|
"loss": 0.585, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 3.9267725944427414, |
|
"learning_rate": 1e-05, |
|
"loss": 0.4444, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"grad_norm": 1.359630389285586, |
|
"learning_rate": 1.5e-05, |
|
"loss": 0.1918, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 0.8675984069699776, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0255, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"grad_norm": 1.3166063968433408, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.0094, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"grad_norm": 1.4298321029000574, |
|
"learning_rate": 3e-05, |
|
"loss": 0.0066, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"grad_norm": 0.39016346214571596, |
|
"learning_rate": 3.5e-05, |
|
"loss": 0.0054, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"grad_norm": 0.2790201303316679, |
|
"learning_rate": 4e-05, |
|
"loss": 0.0062, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 22.5, |
|
"grad_norm": 0.08501619599523778, |
|
"learning_rate": 4.5e-05, |
|
"loss": 0.004, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"grad_norm": 0.062067934755517584, |
|
"learning_rate": 5e-05, |
|
"loss": 0.0029, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 27.5, |
|
"grad_norm": 0.052666305930186126, |
|
"learning_rate": 4.9996924362330555e-05, |
|
"loss": 0.0028, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"grad_norm": 0.03938136111188392, |
|
"learning_rate": 4.998769829017084e-05, |
|
"loss": 0.0028, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 32.5, |
|
"grad_norm": 0.03394672644933696, |
|
"learning_rate": 4.997232430583686e-05, |
|
"loss": 0.0028, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"grad_norm": 0.037661243334742, |
|
"learning_rate": 4.995080661242243e-05, |
|
"loss": 0.0028, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 37.5, |
|
"grad_norm": 0.039184906370568255, |
|
"learning_rate": 4.992315109265007e-05, |
|
"loss": 0.0027, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"grad_norm": 0.04435589843124744, |
|
"learning_rate": 4.988936530726276e-05, |
|
"loss": 0.0028, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 42.5, |
|
"grad_norm": 0.02657151726999846, |
|
"learning_rate": 4.984945849295686e-05, |
|
"loss": 0.0026, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"grad_norm": 0.03472385835097059, |
|
"learning_rate": 4.980344155985695e-05, |
|
"loss": 0.0027, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 47.5, |
|
"grad_norm": 0.031070975559645563, |
|
"learning_rate": 4.975132708853304e-05, |
|
"loss": 0.0026, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"grad_norm": 0.03730805984645713, |
|
"learning_rate": 4.9693129326561254e-05, |
|
"loss": 0.0027, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 52.5, |
|
"grad_norm": 0.030401417138752773, |
|
"learning_rate": 4.96288641846286e-05, |
|
"loss": 0.0027, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 55.0, |
|
"grad_norm": 0.030056720425659458, |
|
"learning_rate": 4.955854923218321e-05, |
|
"loss": 0.0027, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 57.5, |
|
"grad_norm": 0.02714242610959012, |
|
"learning_rate": 4.948220369263099e-05, |
|
"loss": 0.0025, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"grad_norm": 0.021468304046160087, |
|
"learning_rate": 4.939984843808013e-05, |
|
"loss": 0.0026, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 62.5, |
|
"grad_norm": 0.027579476441235425, |
|
"learning_rate": 4.931150598363494e-05, |
|
"loss": 0.0027, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 65.0, |
|
"grad_norm": 0.041471384473431336, |
|
"learning_rate": 4.921720048124035e-05, |
|
"loss": 0.0028, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 67.5, |
|
"grad_norm": 0.030743428074535287, |
|
"learning_rate": 4.9116957713079084e-05, |
|
"loss": 0.0027, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"grad_norm": 0.03200739761669791, |
|
"learning_rate": 4.901080508452305e-05, |
|
"loss": 0.0027, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 72.5, |
|
"grad_norm": 0.02266656589663142, |
|
"learning_rate": 4.889877161664096e-05, |
|
"loss": 0.0027, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 75.0, |
|
"grad_norm": 0.03858968870626217, |
|
"learning_rate": 4.878088793826428e-05, |
|
"loss": 0.0026, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 77.5, |
|
"grad_norm": 0.0436229252420695, |
|
"learning_rate": 4.865718627761363e-05, |
|
"loss": 0.0027, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"grad_norm": 0.021505547895787954, |
|
"learning_rate": 4.8527700453487873e-05, |
|
"loss": 0.0026, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 82.5, |
|
"grad_norm": 0.020985544413640905, |
|
"learning_rate": 4.839246586601841e-05, |
|
"loss": 0.0027, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 85.0, |
|
"grad_norm": 0.023894609006279273, |
|
"learning_rate": 4.825151948699116e-05, |
|
"loss": 0.0028, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 87.5, |
|
"grad_norm": 0.02603189523521318, |
|
"learning_rate": 4.8104899849738795e-05, |
|
"loss": 0.0026, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 90.0, |
|
"grad_norm": 0.033072587670050196, |
|
"learning_rate": 4.795264703860616e-05, |
|
"loss": 0.0027, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 92.5, |
|
"grad_norm": 0.029455707311945873, |
|
"learning_rate": 4.779480267799158e-05, |
|
"loss": 0.0027, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 95.0, |
|
"grad_norm": 0.031066744462249135, |
|
"learning_rate": 4.763140992096718e-05, |
|
"loss": 0.0027, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 97.5, |
|
"grad_norm": 0.03179467575788789, |
|
"learning_rate": 4.7462513437481266e-05, |
|
"loss": 0.0027, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"grad_norm": 0.04391757622955937, |
|
"learning_rate": 4.7288159402146e-05, |
|
"loss": 0.0027, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 102.5, |
|
"grad_norm": 0.029394885252518814, |
|
"learning_rate": 4.7108395481613736e-05, |
|
"loss": 0.0028, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 105.0, |
|
"grad_norm": 0.027320800240786685, |
|
"learning_rate": 4.692327082154542e-05, |
|
"loss": 0.0027, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 107.5, |
|
"grad_norm": 0.030873484414188097, |
|
"learning_rate": 4.6732836033174634e-05, |
|
"loss": 0.0027, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 110.0, |
|
"grad_norm": 0.048697733723929194, |
|
"learning_rate": 4.653714317947105e-05, |
|
"loss": 0.0027, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 112.5, |
|
"grad_norm": 0.025217762723189756, |
|
"learning_rate": 4.6336245760906896e-05, |
|
"loss": 0.0026, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 115.0, |
|
"grad_norm": 0.034325242650264584, |
|
"learning_rate": 4.6130198700830455e-05, |
|
"loss": 0.0027, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 117.5, |
|
"grad_norm": 0.0392844047749773, |
|
"learning_rate": 4.591905833045059e-05, |
|
"loss": 0.0027, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"grad_norm": 0.02965200107252582, |
|
"learning_rate": 4.570288237343632e-05, |
|
"loss": 0.0026, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 122.5, |
|
"grad_norm": 0.020210067730727373, |
|
"learning_rate": 4.5481729930135805e-05, |
|
"loss": 0.0026, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 125.0, |
|
"grad_norm": 0.027623657423726022, |
|
"learning_rate": 4.5255661461418854e-05, |
|
"loss": 0.0026, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 127.5, |
|
"grad_norm": 0.023255879069235798, |
|
"learning_rate": 4.502473877214754e-05, |
|
"loss": 0.0027, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 130.0, |
|
"grad_norm": 0.014768241964023569, |
|
"learning_rate": 4.478902499427933e-05, |
|
"loss": 0.0027, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 132.5, |
|
"grad_norm": 0.03668790355624439, |
|
"learning_rate": 4.454858456960754e-05, |
|
"loss": 0.0027, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 135.0, |
|
"grad_norm": 0.028968111779672595, |
|
"learning_rate": 4.430348323214347e-05, |
|
"loss": 0.0026, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 137.5, |
|
"grad_norm": 0.02287695653834384, |
|
"learning_rate": 4.4053787990145465e-05, |
|
"loss": 0.0027, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 140.0, |
|
"grad_norm": 0.014260556347040123, |
|
"learning_rate": 4.379956710779951e-05, |
|
"loss": 0.0027, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 142.5, |
|
"grad_norm": 0.029771289364298407, |
|
"learning_rate": 4.3540890086556435e-05, |
|
"loss": 0.0026, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 145.0, |
|
"grad_norm": 0.03788961751665883, |
|
"learning_rate": 4.327782764613099e-05, |
|
"loss": 0.0027, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 147.5, |
|
"grad_norm": 0.03971119187487509, |
|
"learning_rate": 4.301045170516773e-05, |
|
"loss": 0.0026, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 150.0, |
|
"grad_norm": 0.03370808514177945, |
|
"learning_rate": 4.2738835361579175e-05, |
|
"loss": 0.0026, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 152.5, |
|
"grad_norm": 0.04042652697228835, |
|
"learning_rate": 4.2463052872561584e-05, |
|
"loss": 0.0027, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 155.0, |
|
"grad_norm": 0.03521656241677526, |
|
"learning_rate": 4.2183179634293794e-05, |
|
"loss": 0.0028, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 157.5, |
|
"grad_norm": 0.03092781528226213, |
|
"learning_rate": 4.1899292161324627e-05, |
|
"loss": 0.0026, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 160.0, |
|
"grad_norm": 0.020310997173564212, |
|
"learning_rate": 4.1611468065654586e-05, |
|
"loss": 0.0026, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 162.5, |
|
"grad_norm": 0.015790567180341695, |
|
"learning_rate": 4.1319786035517534e-05, |
|
"loss": 0.0027, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 165.0, |
|
"grad_norm": 0.0194147951616345, |
|
"learning_rate": 4.1024325813868065e-05, |
|
"loss": 0.0027, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 167.5, |
|
"grad_norm": 0.030730482667734805, |
|
"learning_rate": 4.072516817658065e-05, |
|
"loss": 0.0027, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 170.0, |
|
"grad_norm": 0.029837731953442564, |
|
"learning_rate": 4.0422394910366236e-05, |
|
"loss": 0.0026, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 172.5, |
|
"grad_norm": 0.02492166177013941, |
|
"learning_rate": 4.0116088790412645e-05, |
|
"loss": 0.0026, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 175.0, |
|
"grad_norm": 0.022790129612325603, |
|
"learning_rate": 3.980633355775461e-05, |
|
"loss": 0.0027, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 177.5, |
|
"grad_norm": 0.0367286949474086, |
|
"learning_rate": 3.949321389637986e-05, |
|
"loss": 0.0027, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 180.0, |
|
"grad_norm": 0.026449210914021956, |
|
"learning_rate": 3.917681541007734e-05, |
|
"loss": 0.0025, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 182.5, |
|
"grad_norm": 0.025906355217603764, |
|
"learning_rate": 3.885722459903399e-05, |
|
"loss": 0.0027, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 185.0, |
|
"grad_norm": 0.013510506972149507, |
|
"learning_rate": 3.853452883618644e-05, |
|
"loss": 0.0027, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 187.5, |
|
"grad_norm": 0.02126782273762181, |
|
"learning_rate": 3.8208816343334156e-05, |
|
"loss": 0.0027, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 190.0, |
|
"grad_norm": 0.02778025887186039, |
|
"learning_rate": 3.788017616702048e-05, |
|
"loss": 0.0027, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 192.5, |
|
"grad_norm": 0.02023158670094667, |
|
"learning_rate": 3.754869815418815e-05, |
|
"loss": 0.0026, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 195.0, |
|
"grad_norm": 0.031240526748051056, |
|
"learning_rate": 3.721447292761609e-05, |
|
"loss": 0.0027, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 197.5, |
|
"grad_norm": 0.02268049791605829, |
|
"learning_rate": 3.687759186114403e-05, |
|
"loss": 0.0026, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 200.0, |
|
"grad_norm": 0.02998474683024178, |
|
"learning_rate": 3.6538147054691817e-05, |
|
"loss": 0.0027, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 202.5, |
|
"grad_norm": 0.026430986104193692, |
|
"learning_rate": 3.619623130908018e-05, |
|
"loss": 0.0027, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 205.0, |
|
"grad_norm": 0.010605276850508178, |
|
"learning_rate": 3.5851938100659964e-05, |
|
"loss": 0.0027, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 207.5, |
|
"grad_norm": 0.02068209169464807, |
|
"learning_rate": 3.550536155575662e-05, |
|
"loss": 0.0027, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 210.0, |
|
"grad_norm": 0.03300339570363238, |
|
"learning_rate": 3.515659642493697e-05, |
|
"loss": 0.0026, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 212.5, |
|
"grad_norm": 0.026016657903437497, |
|
"learning_rate": 3.480573805710538e-05, |
|
"loss": 0.0027, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 215.0, |
|
"grad_norm": 0.024770792616580162, |
|
"learning_rate": 3.4452882373436316e-05, |
|
"loss": 0.0026, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 217.5, |
|
"grad_norm": 0.0265444448484458, |
|
"learning_rate": 3.4098125841150466e-05, |
|
"loss": 0.0026, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 220.0, |
|
"grad_norm": 0.01924921498743014, |
|
"learning_rate": 3.37415654471415e-05, |
|
"loss": 0.0026, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 222.5, |
|
"grad_norm": 0.027603952267821173, |
|
"learning_rate": 3.3383298671460944e-05, |
|
"loss": 0.0026, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 225.0, |
|
"grad_norm": 0.0254709928199966, |
|
"learning_rate": 3.3023423460667985e-05, |
|
"loss": 0.0026, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 227.5, |
|
"grad_norm": 0.031183288661866662, |
|
"learning_rate": 3.2662038201051914e-05, |
|
"loss": 0.0027, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 230.0, |
|
"grad_norm": 0.03425285898227886, |
|
"learning_rate": 3.2299241691734304e-05, |
|
"loss": 0.0026, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 232.5, |
|
"grad_norm": 0.028752151160391863, |
|
"learning_rate": 3.1935133117658306e-05, |
|
"loss": 0.0027, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 235.0, |
|
"grad_norm": 0.028976330470022123, |
|
"learning_rate": 3.156981202247248e-05, |
|
"loss": 0.0026, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 237.5, |
|
"grad_norm": 0.026920539664030775, |
|
"learning_rate": 3.1203378281316515e-05, |
|
"loss": 0.0026, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 240.0, |
|
"grad_norm": 0.020989002709489214, |
|
"learning_rate": 3.0835932073516444e-05, |
|
"loss": 0.0027, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 242.5, |
|
"grad_norm": 0.024523095658861207, |
|
"learning_rate": 3.0467573855196558e-05, |
|
"loss": 0.0027, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 245.0, |
|
"grad_norm": 0.015580401724056997, |
|
"learning_rate": 3.0098404331815695e-05, |
|
"loss": 0.0027, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 247.5, |
|
"grad_norm": 0.022996840296875175, |
|
"learning_rate": 2.9728524430635417e-05, |
|
"loss": 0.0027, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 250.0, |
|
"grad_norm": 0.021854877708908236, |
|
"learning_rate": 2.9358035273127483e-05, |
|
"loss": 0.0026, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 252.5, |
|
"grad_norm": 0.025003129400125675, |
|
"learning_rate": 2.8987038147328238e-05, |
|
"loss": 0.0026, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 255.0, |
|
"grad_norm": 0.02759119149124416, |
|
"learning_rate": 2.86156344801475e-05, |
|
"loss": 0.0026, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 257.5, |
|
"grad_norm": 0.024349103530368846, |
|
"learning_rate": 2.824392580963944e-05, |
|
"loss": 0.0027, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 260.0, |
|
"grad_norm": 0.02918200596645593, |
|
"learning_rate": 2.787201375724307e-05, |
|
"loss": 0.0027, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 262.5, |
|
"grad_norm": 0.017661985537795705, |
|
"learning_rate": 2.7500000000000004e-05, |
|
"loss": 0.0026, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 265.0, |
|
"grad_norm": 0.02880881596677379, |
|
"learning_rate": 2.7127986242756936e-05, |
|
"loss": 0.0026, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 267.5, |
|
"grad_norm": 0.04090166133876468, |
|
"learning_rate": 2.6756074190360563e-05, |
|
"loss": 0.0028, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 270.0, |
|
"grad_norm": 0.022838636369005745, |
|
"learning_rate": 2.63843655198525e-05, |
|
"loss": 0.0027, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 272.5, |
|
"grad_norm": 0.03417808915995595, |
|
"learning_rate": 2.6012961852671767e-05, |
|
"loss": 0.0026, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 275.0, |
|
"grad_norm": 0.02391514801170413, |
|
"learning_rate": 2.5641964726872526e-05, |
|
"loss": 0.0026, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 277.5, |
|
"grad_norm": 0.027050966559574396, |
|
"learning_rate": 2.527147556936459e-05, |
|
"loss": 0.0027, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 280.0, |
|
"grad_norm": 0.029392338368449532, |
|
"learning_rate": 2.4901595668184314e-05, |
|
"loss": 0.0027, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 282.5, |
|
"grad_norm": 0.03213320609449678, |
|
"learning_rate": 2.453242614480345e-05, |
|
"loss": 0.0027, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 285.0, |
|
"grad_norm": 0.017580210273689653, |
|
"learning_rate": 2.416406792648355e-05, |
|
"loss": 0.0027, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 287.5, |
|
"grad_norm": 0.025395175371267904, |
|
"learning_rate": 2.3796621718683487e-05, |
|
"loss": 0.0027, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 290.0, |
|
"grad_norm": 0.02830946439085919, |
|
"learning_rate": 2.3430187977527533e-05, |
|
"loss": 0.0027, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 292.5, |
|
"grad_norm": 0.04065072947108163, |
|
"learning_rate": 2.3064866882341696e-05, |
|
"loss": 0.0027, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 295.0, |
|
"grad_norm": 0.028524952343665373, |
|
"learning_rate": 2.27007583082657e-05, |
|
"loss": 0.0026, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 297.5, |
|
"grad_norm": 0.0284550999234347, |
|
"learning_rate": 2.233796179894809e-05, |
|
"loss": 0.0026, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 300.0, |
|
"grad_norm": 0.03018543474602915, |
|
"learning_rate": 2.1976576539332024e-05, |
|
"loss": 0.0026, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 302.5, |
|
"grad_norm": 0.017475284356752212, |
|
"learning_rate": 2.1616701328539057e-05, |
|
"loss": 0.0026, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 305.0, |
|
"grad_norm": 0.02381743267091071, |
|
"learning_rate": 2.1258434552858502e-05, |
|
"loss": 0.0026, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 307.5, |
|
"grad_norm": 0.030282727334382083, |
|
"learning_rate": 2.090187415884955e-05, |
|
"loss": 0.0027, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 310.0, |
|
"grad_norm": 0.02239088576419377, |
|
"learning_rate": 2.054711762656369e-05, |
|
"loss": 0.0026, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 312.5, |
|
"grad_norm": 0.020772310557568244, |
|
"learning_rate": 2.0194261942894628e-05, |
|
"loss": 0.0026, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 315.0, |
|
"grad_norm": 0.03490126977201953, |
|
"learning_rate": 1.984340357506303e-05, |
|
"loss": 0.0026, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 317.5, |
|
"grad_norm": 0.015968447976387346, |
|
"learning_rate": 1.949463844424338e-05, |
|
"loss": 0.0026, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 320.0, |
|
"grad_norm": 0.023986554939275363, |
|
"learning_rate": 1.914806189934003e-05, |
|
"loss": 0.0028, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 322.5, |
|
"grad_norm": 0.03420061182554338, |
|
"learning_rate": 1.8803768690919832e-05, |
|
"loss": 0.0027, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 325.0, |
|
"grad_norm": 0.01792951575408546, |
|
"learning_rate": 1.8461852945308196e-05, |
|
"loss": 0.0027, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 327.5, |
|
"grad_norm": 0.023802622532695907, |
|
"learning_rate": 1.8122408138855974e-05, |
|
"loss": 0.0026, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 330.0, |
|
"grad_norm": 0.023316524889500846, |
|
"learning_rate": 1.778552707238391e-05, |
|
"loss": 0.0026, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 332.5, |
|
"grad_norm": 0.013376585581553524, |
|
"learning_rate": 1.7451301845811857e-05, |
|
"loss": 0.0026, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 335.0, |
|
"grad_norm": 0.019080334175566694, |
|
"learning_rate": 1.711982383297953e-05, |
|
"loss": 0.0026, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 337.5, |
|
"grad_norm": 0.01654436855407368, |
|
"learning_rate": 1.6791183656665846e-05, |
|
"loss": 0.0026, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 340.0, |
|
"grad_norm": 0.022278657387787754, |
|
"learning_rate": 1.6465471163813574e-05, |
|
"loss": 0.0027, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 342.5, |
|
"grad_norm": 0.015375948509190046, |
|
"learning_rate": 1.6142775400966023e-05, |
|
"loss": 0.0027, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 345.0, |
|
"grad_norm": 0.01684675867751184, |
|
"learning_rate": 1.582318458992267e-05, |
|
"loss": 0.0027, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 347.5, |
|
"grad_norm": 0.032002543536054376, |
|
"learning_rate": 1.5506786103620146e-05, |
|
"loss": 0.0027, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 350.0, |
|
"grad_norm": 0.021043035856906086, |
|
"learning_rate": 1.5193666442245402e-05, |
|
"loss": 0.0027, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 352.5, |
|
"grad_norm": 0.01601931955071107, |
|
"learning_rate": 1.4883911209587368e-05, |
|
"loss": 0.0026, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 355.0, |
|
"grad_norm": 0.02601502051924795, |
|
"learning_rate": 1.4577605089633773e-05, |
|
"loss": 0.0026, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 357.5, |
|
"grad_norm": 0.02710347701415514, |
|
"learning_rate": 1.427483182341936e-05, |
|
"loss": 0.0026, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 360.0, |
|
"grad_norm": 0.03320644743977735, |
|
"learning_rate": 1.3975674186131937e-05, |
|
"loss": 0.0027, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 362.5, |
|
"grad_norm": 0.01591474795789328, |
|
"learning_rate": 1.3680213964482475e-05, |
|
"loss": 0.0027, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 365.0, |
|
"grad_norm": 0.026155373607272086, |
|
"learning_rate": 1.3388531934345416e-05, |
|
"loss": 0.0027, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 367.5, |
|
"grad_norm": 0.024271143128982743, |
|
"learning_rate": 1.3100707838675377e-05, |
|
"loss": 0.0026, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 370.0, |
|
"grad_norm": 0.016369678310535946, |
|
"learning_rate": 1.2816820365706206e-05, |
|
"loss": 0.0027, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 372.5, |
|
"grad_norm": 0.016955558044549856, |
|
"learning_rate": 1.2536947127438415e-05, |
|
"loss": 0.0026, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 375.0, |
|
"grad_norm": 0.02239783849742695, |
|
"learning_rate": 1.2261164638420832e-05, |
|
"loss": 0.0026, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 377.5, |
|
"grad_norm": 0.029924950679270815, |
|
"learning_rate": 1.198954829483227e-05, |
|
"loss": 0.0027, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 380.0, |
|
"grad_norm": 0.028621427234193805, |
|
"learning_rate": 1.1722172353869008e-05, |
|
"loss": 0.0025, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 382.5, |
|
"grad_norm": 0.02561136953891063, |
|
"learning_rate": 1.1459109913443567e-05, |
|
"loss": 0.0026, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 385.0, |
|
"grad_norm": 0.028256041486332584, |
|
"learning_rate": 1.12004328922005e-05, |
|
"loss": 0.0027, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 387.5, |
|
"grad_norm": 0.029729889846126924, |
|
"learning_rate": 1.094621200985454e-05, |
|
"loss": 0.0027, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 390.0, |
|
"grad_norm": 0.014383243995550028, |
|
"learning_rate": 1.0696516767856546e-05, |
|
"loss": 0.0026, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 392.5, |
|
"grad_norm": 0.015879736540587025, |
|
"learning_rate": 1.0451415430392474e-05, |
|
"loss": 0.0026, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 395.0, |
|
"grad_norm": 0.024629301532455877, |
|
"learning_rate": 1.0210975005720677e-05, |
|
"loss": 0.0026, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 397.5, |
|
"grad_norm": 0.029707059150893382, |
|
"learning_rate": 9.975261227852472e-06, |
|
"loss": 0.0027, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 400.0, |
|
"grad_norm": 0.0318537391592188, |
|
"learning_rate": 9.744338538581147e-06, |
|
"loss": 0.0027, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 402.5, |
|
"grad_norm": 0.017706015342844352, |
|
"learning_rate": 9.518270069864195e-06, |
|
"loss": 0.0027, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 405.0, |
|
"grad_norm": 0.01724610650030409, |
|
"learning_rate": 9.297117626563687e-06, |
|
"loss": 0.0026, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 407.5, |
|
"grad_norm": 0.018957851028085164, |
|
"learning_rate": 9.080941669549423e-06, |
|
"loss": 0.0026, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 410.0, |
|
"grad_norm": 0.025418915769378568, |
|
"learning_rate": 8.86980129916955e-06, |
|
"loss": 0.0026, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 412.5, |
|
"grad_norm": 0.01583223922193541, |
|
"learning_rate": 8.663754239093109e-06, |
|
"loss": 0.0027, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 415.0, |
|
"grad_norm": 0.027164074218547888, |
|
"learning_rate": 8.462856820528952e-06, |
|
"loss": 0.0026, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 417.5, |
|
"grad_norm": 0.028647280916262126, |
|
"learning_rate": 8.26716396682537e-06, |
|
"loss": 0.0027, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 420.0, |
|
"grad_norm": 0.019553123738546497, |
|
"learning_rate": 8.076729178454588e-06, |
|
"loss": 0.0026, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 422.5, |
|
"grad_norm": 0.025379025955769637, |
|
"learning_rate": 7.89160451838626e-06, |
|
"loss": 0.0027, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 425.0, |
|
"grad_norm": 0.01750943138582632, |
|
"learning_rate": 7.711840597853998e-06, |
|
"loss": 0.0026, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 427.5, |
|
"grad_norm": 0.02520822601658372, |
|
"learning_rate": 7.537486562518735e-06, |
|
"loss": 0.0026, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 430.0, |
|
"grad_norm": 0.01993484997782361, |
|
"learning_rate": 7.368590079032822e-06, |
|
"loss": 0.0026, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 432.5, |
|
"grad_norm": 0.019158523070010834, |
|
"learning_rate": 7.205197322008425e-06, |
|
"loss": 0.0027, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 435.0, |
|
"grad_norm": 0.015239850907125138, |
|
"learning_rate": 7.047352961393844e-06, |
|
"loss": 0.0027, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 437.5, |
|
"grad_norm": 0.020559758942860218, |
|
"learning_rate": 6.8951001502612065e-06, |
|
"loss": 0.0027, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 440.0, |
|
"grad_norm": 0.02381279446911383, |
|
"learning_rate": 6.748480513008844e-06, |
|
"loss": 0.0025, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 442.5, |
|
"grad_norm": 0.014589626300082677, |
|
"learning_rate": 6.607534133981594e-06, |
|
"loss": 0.0026, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 445.0, |
|
"grad_norm": 0.026727121558605703, |
|
"learning_rate": 6.472299546512134e-06, |
|
"loss": 0.0026, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 447.5, |
|
"grad_norm": 0.01580348324753083, |
|
"learning_rate": 6.342813722386374e-06, |
|
"loss": 0.0027, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 450.0, |
|
"grad_norm": 0.02334311857794958, |
|
"learning_rate": 6.219112061735721e-06, |
|
"loss": 0.0026, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 452.5, |
|
"grad_norm": 0.022995449553396018, |
|
"learning_rate": 6.1012283833590465e-06, |
|
"loss": 0.0026, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 455.0, |
|
"grad_norm": 0.03391053587500504, |
|
"learning_rate": 5.989194915476954e-06, |
|
"loss": 0.0027, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 457.5, |
|
"grad_norm": 0.0281194121062883, |
|
"learning_rate": 5.883042286920918e-06, |
|
"loss": 0.0025, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 460.0, |
|
"grad_norm": 0.018983581090637, |
|
"learning_rate": 5.782799518759658e-06, |
|
"loss": 0.0025, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 462.5, |
|
"grad_norm": 0.024476355758836905, |
|
"learning_rate": 5.688494016365067e-06, |
|
"loss": 0.0026, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 465.0, |
|
"grad_norm": 0.023061048781595993, |
|
"learning_rate": 5.600151561919871e-06, |
|
"loss": 0.0027, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 467.5, |
|
"grad_norm": 0.012496893287463113, |
|
"learning_rate": 5.517796307369017e-06, |
|
"loss": 0.0026, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 470.0, |
|
"grad_norm": 0.01991683594417727, |
|
"learning_rate": 5.44145076781679e-06, |
|
"loss": 0.0026, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 472.5, |
|
"grad_norm": 0.028037001206555136, |
|
"learning_rate": 5.371135815371398e-06, |
|
"loss": 0.0027, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 475.0, |
|
"grad_norm": 0.026283232329898983, |
|
"learning_rate": 5.3068706734387484e-06, |
|
"loss": 0.0026, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 477.5, |
|
"grad_norm": 0.026815406305102027, |
|
"learning_rate": 5.248672911466959e-06, |
|
"loss": 0.0026, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 480.0, |
|
"grad_norm": 0.01847886314594556, |
|
"learning_rate": 5.196558440143059e-06, |
|
"loss": 0.0025, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 482.5, |
|
"grad_norm": 0.019370493214584474, |
|
"learning_rate": 5.150541507043143e-06, |
|
"loss": 0.0027, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 485.0, |
|
"grad_norm": 0.021779585772489417, |
|
"learning_rate": 5.110634692737244e-06, |
|
"loss": 0.0026, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 487.5, |
|
"grad_norm": 0.02541037015533656, |
|
"learning_rate": 5.0768489073499295e-06, |
|
"loss": 0.0027, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 490.0, |
|
"grad_norm": 0.02945361459676501, |
|
"learning_rate": 5.049193387577574e-06, |
|
"loss": 0.0026, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 492.5, |
|
"grad_norm": 0.017922772556181753, |
|
"learning_rate": 5.027675694163144e-06, |
|
"loss": 0.0028, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 495.0, |
|
"grad_norm": 0.023023048365518207, |
|
"learning_rate": 5.012301709829164e-06, |
|
"loss": 0.0026, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 497.5, |
|
"grad_norm": 0.020646776127788924, |
|
"learning_rate": 5.003075637669448e-06, |
|
"loss": 0.0027, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 500.0, |
|
"grad_norm": 0.0148034159233786, |
|
"learning_rate": 5e-06, |
|
"loss": 0.0027, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 500.0, |
|
"step": 1000, |
|
"total_flos": 28457722773504.0, |
|
"train_loss": 0.008928275344893336, |
|
"train_runtime": 1701.2794, |
|
"train_samples_per_second": 18.809, |
|
"train_steps_per_second": 0.588 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 1000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 500, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 28457722773504.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|