{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9999059177721329, "eval_steps": 20, "global_step": 5314, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0037632891146862357, "grad_norm": 0.15685315430164337, "learning_rate": 0.0002, "loss": 0.4618, "step": 20 }, { "epoch": 0.0075265782293724715, "grad_norm": 0.10632659494876862, "learning_rate": 0.0002, "loss": 0.2985, "step": 40 }, { "epoch": 0.011289867344058707, "grad_norm": 0.12228264659643173, "learning_rate": 0.0002, "loss": 0.2689, "step": 60 }, { "epoch": 0.015053156458744943, "grad_norm": 0.09272768348455429, "learning_rate": 0.0002, "loss": 0.2249, "step": 80 }, { "epoch": 0.01881644557343118, "grad_norm": 0.11586301028728485, "learning_rate": 0.0002, "loss": 0.23, "step": 100 }, { "epoch": 0.022579734688117414, "grad_norm": 0.08657937496900558, "learning_rate": 0.0002, "loss": 0.2189, "step": 120 }, { "epoch": 0.02634302380280365, "grad_norm": 0.08374184370040894, "learning_rate": 0.0002, "loss": 0.1987, "step": 140 }, { "epoch": 0.030106312917489886, "grad_norm": 0.08528616279363632, "learning_rate": 0.0002, "loss": 0.2065, "step": 160 }, { "epoch": 0.03386960203217612, "grad_norm": 0.17550894618034363, "learning_rate": 0.0002, "loss": 0.1904, "step": 180 }, { "epoch": 0.03763289114686236, "grad_norm": 0.09562012553215027, "learning_rate": 0.0002, "loss": 0.1824, "step": 200 }, { "epoch": 0.04139618026154859, "grad_norm": 0.12333519756793976, "learning_rate": 0.0002, "loss": 0.1702, "step": 220 }, { "epoch": 0.04515946937623483, "grad_norm": 0.11404936760663986, "learning_rate": 0.0002, "loss": 0.18, "step": 240 }, { "epoch": 0.048922758490921064, "grad_norm": 0.08656694740056992, "learning_rate": 0.0002, "loss": 0.1749, "step": 260 }, { "epoch": 0.0526860476056073, "grad_norm": 0.09797225147485733, "learning_rate": 0.0002, "loss": 0.1731, "step": 280 }, { "epoch": 0.056449336720293536, "grad_norm": 0.09765412658452988, "learning_rate": 0.0002, "loss": 0.1578, "step": 300 }, { "epoch": 0.06021262583497977, "grad_norm": 0.07540671527385712, "learning_rate": 0.0002, "loss": 0.1693, "step": 320 }, { "epoch": 0.06397591494966601, "grad_norm": 0.08590289205312729, "learning_rate": 0.0002, "loss": 0.1607, "step": 340 }, { "epoch": 0.06773920406435224, "grad_norm": 0.09767664223909378, "learning_rate": 0.0002, "loss": 0.1419, "step": 360 }, { "epoch": 0.07150249317903848, "grad_norm": 0.10479151457548141, "learning_rate": 0.0002, "loss": 0.1629, "step": 380 }, { "epoch": 0.07526578229372471, "grad_norm": 0.08791118115186691, "learning_rate": 0.0002, "loss": 0.1484, "step": 400 }, { "epoch": 0.07902907140841095, "grad_norm": 0.10221686214208603, "learning_rate": 0.0002, "loss": 0.1499, "step": 420 }, { "epoch": 0.08279236052309719, "grad_norm": 0.09131903946399689, "learning_rate": 0.0002, "loss": 0.1485, "step": 440 }, { "epoch": 0.08655564963778342, "grad_norm": 0.10372031480073929, "learning_rate": 0.0002, "loss": 0.1441, "step": 460 }, { "epoch": 0.09031893875246966, "grad_norm": 0.09649350494146347, "learning_rate": 0.0002, "loss": 0.147, "step": 480 }, { "epoch": 0.0940822278671559, "grad_norm": 0.09961670637130737, "learning_rate": 0.0002, "loss": 0.1465, "step": 500 }, { "epoch": 0.09784551698184213, "grad_norm": 0.08490657806396484, "learning_rate": 0.0002, "loss": 0.132, "step": 520 }, { "epoch": 0.10160880609652836, "grad_norm": 0.08765380829572678, "learning_rate": 0.0002, "loss": 0.1283, "step": 540 }, { "epoch": 0.1053720952112146, "grad_norm": 0.09319768846035004, "learning_rate": 0.0002, "loss": 0.1382, "step": 560 }, { "epoch": 0.10913538432590084, "grad_norm": 0.08941628783941269, "learning_rate": 0.0002, "loss": 0.1335, "step": 580 }, { "epoch": 0.11289867344058707, "grad_norm": 0.0971933901309967, "learning_rate": 0.0002, "loss": 0.137, "step": 600 }, { "epoch": 0.11666196255527331, "grad_norm": 0.07488075643777847, "learning_rate": 0.0002, "loss": 0.1283, "step": 620 }, { "epoch": 0.12042525166995954, "grad_norm": 0.08711710572242737, "learning_rate": 0.0002, "loss": 0.1292, "step": 640 }, { "epoch": 0.12418854078464578, "grad_norm": 0.08043856918811798, "learning_rate": 0.0002, "loss": 0.1269, "step": 660 }, { "epoch": 0.12795182989933201, "grad_norm": 0.07097792625427246, "learning_rate": 0.0002, "loss": 0.1364, "step": 680 }, { "epoch": 0.13171511901401825, "grad_norm": 0.0742156058549881, "learning_rate": 0.0002, "loss": 0.1277, "step": 700 }, { "epoch": 0.1354784081287045, "grad_norm": 0.09956187009811401, "learning_rate": 0.0002, "loss": 0.1314, "step": 720 }, { "epoch": 0.13924169724339072, "grad_norm": 0.08661571145057678, "learning_rate": 0.0002, "loss": 0.124, "step": 740 }, { "epoch": 0.14300498635807696, "grad_norm": 0.08070897310972214, "learning_rate": 0.0002, "loss": 0.1147, "step": 760 }, { "epoch": 0.1467682754727632, "grad_norm": 0.09516704082489014, "learning_rate": 0.0002, "loss": 0.1168, "step": 780 }, { "epoch": 0.15053156458744943, "grad_norm": 0.09076276421546936, "learning_rate": 0.0002, "loss": 0.1178, "step": 800 }, { "epoch": 0.15429485370213566, "grad_norm": 0.0896017774939537, "learning_rate": 0.0002, "loss": 0.1192, "step": 820 }, { "epoch": 0.1580581428168219, "grad_norm": 0.07477965205907822, "learning_rate": 0.0002, "loss": 0.1127, "step": 840 }, { "epoch": 0.16182143193150814, "grad_norm": 0.08405464887619019, "learning_rate": 0.0002, "loss": 0.119, "step": 860 }, { "epoch": 0.16558472104619437, "grad_norm": 0.07539790868759155, "learning_rate": 0.0002, "loss": 0.1213, "step": 880 }, { "epoch": 0.1693480101608806, "grad_norm": 0.08806908130645752, "learning_rate": 0.0002, "loss": 0.1192, "step": 900 }, { "epoch": 0.17311129927556684, "grad_norm": 0.08064749836921692, "learning_rate": 0.0002, "loss": 0.1141, "step": 920 }, { "epoch": 0.17687458839025308, "grad_norm": 0.09680119901895523, "learning_rate": 0.0002, "loss": 0.1042, "step": 940 }, { "epoch": 0.18063787750493931, "grad_norm": 0.09092500060796738, "learning_rate": 0.0002, "loss": 0.1067, "step": 960 }, { "epoch": 0.18440116661962555, "grad_norm": 0.07870171219110489, "learning_rate": 0.0002, "loss": 0.1101, "step": 980 }, { "epoch": 0.1881644557343118, "grad_norm": 0.06842092424631119, "learning_rate": 0.0002, "loss": 0.1045, "step": 1000 }, { "epoch": 0.19192774484899802, "grad_norm": 0.08229291439056396, "learning_rate": 0.0002, "loss": 0.114, "step": 1020 }, { "epoch": 0.19569103396368426, "grad_norm": 0.07617371529340744, "learning_rate": 0.0002, "loss": 0.1077, "step": 1040 }, { "epoch": 0.1994543230783705, "grad_norm": 0.09529408067464828, "learning_rate": 0.0002, "loss": 0.101, "step": 1060 }, { "epoch": 0.20321761219305673, "grad_norm": 0.07952335476875305, "learning_rate": 0.0002, "loss": 0.1065, "step": 1080 }, { "epoch": 0.20698090130774296, "grad_norm": 0.07625720649957657, "learning_rate": 0.0002, "loss": 0.0987, "step": 1100 }, { "epoch": 0.2107441904224292, "grad_norm": 0.07906854152679443, "learning_rate": 0.0002, "loss": 0.0986, "step": 1120 }, { "epoch": 0.21450747953711544, "grad_norm": 0.08430849015712738, "learning_rate": 0.0002, "loss": 0.1085, "step": 1140 }, { "epoch": 0.21827076865180167, "grad_norm": 0.0861297994852066, "learning_rate": 0.0002, "loss": 0.1066, "step": 1160 }, { "epoch": 0.2220340577664879, "grad_norm": 0.07576191425323486, "learning_rate": 0.0002, "loss": 0.1051, "step": 1180 }, { "epoch": 0.22579734688117414, "grad_norm": 0.08132428675889969, "learning_rate": 0.0002, "loss": 0.1036, "step": 1200 }, { "epoch": 0.22956063599586038, "grad_norm": 0.06960251182317734, "learning_rate": 0.0002, "loss": 0.1039, "step": 1220 }, { "epoch": 0.23332392511054661, "grad_norm": 0.08230841159820557, "learning_rate": 0.0002, "loss": 0.1046, "step": 1240 }, { "epoch": 0.23708721422523285, "grad_norm": 0.07119760662317276, "learning_rate": 0.0002, "loss": 0.1064, "step": 1260 }, { "epoch": 0.2408505033399191, "grad_norm": 0.06965576857328415, "learning_rate": 0.0002, "loss": 0.0964, "step": 1280 }, { "epoch": 0.24461379245460532, "grad_norm": 0.07533243298530579, "learning_rate": 0.0002, "loss": 0.0973, "step": 1300 }, { "epoch": 0.24837708156929156, "grad_norm": 0.07530753314495087, "learning_rate": 0.0002, "loss": 0.1002, "step": 1320 }, { "epoch": 0.2521403706839778, "grad_norm": 0.0701604038476944, "learning_rate": 0.0002, "loss": 0.1014, "step": 1340 }, { "epoch": 0.25590365979866403, "grad_norm": 0.08768032491207123, "learning_rate": 0.0002, "loss": 0.099, "step": 1360 }, { "epoch": 0.2596669489133503, "grad_norm": 0.0789860337972641, "learning_rate": 0.0002, "loss": 0.0947, "step": 1380 }, { "epoch": 0.2634302380280365, "grad_norm": 0.09132009744644165, "learning_rate": 0.0002, "loss": 0.0975, "step": 1400 }, { "epoch": 0.26719352714272276, "grad_norm": 0.07386859506368637, "learning_rate": 0.0002, "loss": 0.0997, "step": 1420 }, { "epoch": 0.270956816257409, "grad_norm": 0.07243089377880096, "learning_rate": 0.0002, "loss": 0.0956, "step": 1440 }, { "epoch": 0.27472010537209524, "grad_norm": 0.0814971774816513, "learning_rate": 0.0002, "loss": 0.0869, "step": 1460 }, { "epoch": 0.27848339448678144, "grad_norm": 0.07246191054582596, "learning_rate": 0.0002, "loss": 0.0958, "step": 1480 }, { "epoch": 0.2822466836014677, "grad_norm": 0.08997531235218048, "learning_rate": 0.0002, "loss": 0.0875, "step": 1500 }, { "epoch": 0.2860099727161539, "grad_norm": 0.08498572558164597, "learning_rate": 0.0002, "loss": 0.0872, "step": 1520 }, { "epoch": 0.2897732618308402, "grad_norm": 0.08210768550634384, "learning_rate": 0.0002, "loss": 0.085, "step": 1540 }, { "epoch": 0.2935365509455264, "grad_norm": 0.09801402688026428, "learning_rate": 0.0002, "loss": 0.0937, "step": 1560 }, { "epoch": 0.29729984006021265, "grad_norm": 0.07310175150632858, "learning_rate": 0.0002, "loss": 0.0898, "step": 1580 }, { "epoch": 0.30106312917489886, "grad_norm": 0.0679851844906807, "learning_rate": 0.0002, "loss": 0.0853, "step": 1600 }, { "epoch": 0.3048264182895851, "grad_norm": 0.0764717161655426, "learning_rate": 0.0002, "loss": 0.0966, "step": 1620 }, { "epoch": 0.30858970740427133, "grad_norm": 0.0813748836517334, "learning_rate": 0.0002, "loss": 0.0914, "step": 1640 }, { "epoch": 0.3123529965189576, "grad_norm": 0.07560984790325165, "learning_rate": 0.0002, "loss": 0.0889, "step": 1660 }, { "epoch": 0.3161162856336438, "grad_norm": 0.07131631672382355, "learning_rate": 0.0002, "loss": 0.0862, "step": 1680 }, { "epoch": 0.31987957474833006, "grad_norm": 0.07922184467315674, "learning_rate": 0.0002, "loss": 0.0908, "step": 1700 }, { "epoch": 0.3236428638630163, "grad_norm": 0.09043222665786743, "learning_rate": 0.0002, "loss": 0.0887, "step": 1720 }, { "epoch": 0.32740615297770254, "grad_norm": 0.08574160188436508, "learning_rate": 0.0002, "loss": 0.0895, "step": 1740 }, { "epoch": 0.33116944209238874, "grad_norm": 0.0885438472032547, "learning_rate": 0.0002, "loss": 0.0886, "step": 1760 }, { "epoch": 0.334932731207075, "grad_norm": 0.08546678721904755, "learning_rate": 0.0002, "loss": 0.0871, "step": 1780 }, { "epoch": 0.3386960203217612, "grad_norm": 0.0922817662358284, "learning_rate": 0.0002, "loss": 0.0877, "step": 1800 }, { "epoch": 0.3424593094364475, "grad_norm": 0.08879639208316803, "learning_rate": 0.0002, "loss": 0.0877, "step": 1820 }, { "epoch": 0.3462225985511337, "grad_norm": 0.09202056378126144, "learning_rate": 0.0002, "loss": 0.0827, "step": 1840 }, { "epoch": 0.34998588766581995, "grad_norm": 0.0998295396566391, "learning_rate": 0.0002, "loss": 0.0835, "step": 1860 }, { "epoch": 0.35374917678050616, "grad_norm": 0.09490591287612915, "learning_rate": 0.0002, "loss": 0.0862, "step": 1880 }, { "epoch": 0.3575124658951924, "grad_norm": 0.08920489251613617, "learning_rate": 0.0002, "loss": 0.0859, "step": 1900 }, { "epoch": 0.36127575500987863, "grad_norm": 0.0803561583161354, "learning_rate": 0.0002, "loss": 0.081, "step": 1920 }, { "epoch": 0.3650390441245649, "grad_norm": 0.08496900647878647, "learning_rate": 0.0002, "loss": 0.0835, "step": 1940 }, { "epoch": 0.3688023332392511, "grad_norm": 0.08658849447965622, "learning_rate": 0.0002, "loss": 0.0883, "step": 1960 }, { "epoch": 0.37256562235393736, "grad_norm": 0.08497461676597595, "learning_rate": 0.0002, "loss": 0.0811, "step": 1980 }, { "epoch": 0.3763289114686236, "grad_norm": 0.06599749624729156, "learning_rate": 0.0002, "loss": 0.0714, "step": 2000 }, { "epoch": 0.38009220058330984, "grad_norm": 0.07988911867141724, "learning_rate": 0.0002, "loss": 0.0874, "step": 2020 }, { "epoch": 0.38385548969799604, "grad_norm": 0.07323348522186279, "learning_rate": 0.0002, "loss": 0.0809, "step": 2040 }, { "epoch": 0.3876187788126823, "grad_norm": 0.07455869019031525, "learning_rate": 0.0002, "loss": 0.0819, "step": 2060 }, { "epoch": 0.3913820679273685, "grad_norm": 0.08383121341466904, "learning_rate": 0.0002, "loss": 0.0811, "step": 2080 }, { "epoch": 0.3951453570420548, "grad_norm": 0.07616332918405533, "learning_rate": 0.0002, "loss": 0.0802, "step": 2100 }, { "epoch": 0.398908646156741, "grad_norm": 0.08373293280601501, "learning_rate": 0.0002, "loss": 0.0878, "step": 2120 }, { "epoch": 0.40267193527142725, "grad_norm": 0.07459353655576706, "learning_rate": 0.0002, "loss": 0.0806, "step": 2140 }, { "epoch": 0.40643522438611346, "grad_norm": 0.08267400413751602, "learning_rate": 0.0002, "loss": 0.0799, "step": 2160 }, { "epoch": 0.4101985135007997, "grad_norm": 0.07844484597444534, "learning_rate": 0.0002, "loss": 0.0843, "step": 2180 }, { "epoch": 0.41396180261548593, "grad_norm": 0.07998470962047577, "learning_rate": 0.0002, "loss": 0.0821, "step": 2200 }, { "epoch": 0.4177250917301722, "grad_norm": 0.08860679715871811, "learning_rate": 0.0002, "loss": 0.0819, "step": 2220 }, { "epoch": 0.4214883808448584, "grad_norm": 0.07845838367938995, "learning_rate": 0.0002, "loss": 0.0802, "step": 2240 }, { "epoch": 0.42525166995954466, "grad_norm": 0.08801402896642685, "learning_rate": 0.0002, "loss": 0.0768, "step": 2260 }, { "epoch": 0.4290149590742309, "grad_norm": 0.06149598956108093, "learning_rate": 0.0002, "loss": 0.0806, "step": 2280 }, { "epoch": 0.43277824818891714, "grad_norm": 0.07813749462366104, "learning_rate": 0.0002, "loss": 0.0775, "step": 2300 }, { "epoch": 0.43654153730360334, "grad_norm": 0.08169027417898178, "learning_rate": 0.0002, "loss": 0.0774, "step": 2320 }, { "epoch": 0.4403048264182896, "grad_norm": 0.06748262792825699, "learning_rate": 0.0002, "loss": 0.0721, "step": 2340 }, { "epoch": 0.4440681155329758, "grad_norm": 0.07497742772102356, "learning_rate": 0.0002, "loss": 0.0687, "step": 2360 }, { "epoch": 0.4478314046476621, "grad_norm": 0.08851341158151627, "learning_rate": 0.0002, "loss": 0.0809, "step": 2380 }, { "epoch": 0.4515946937623483, "grad_norm": 0.06805267184972763, "learning_rate": 0.0002, "loss": 0.0783, "step": 2400 }, { "epoch": 0.45535798287703455, "grad_norm": 0.07998234033584595, "learning_rate": 0.0002, "loss": 0.0816, "step": 2420 }, { "epoch": 0.45912127199172076, "grad_norm": 0.07358496636152267, "learning_rate": 0.0002, "loss": 0.0715, "step": 2440 }, { "epoch": 0.462884561106407, "grad_norm": 0.08105491101741791, "learning_rate": 0.0002, "loss": 0.076, "step": 2460 }, { "epoch": 0.46664785022109323, "grad_norm": 0.08625109493732452, "learning_rate": 0.0002, "loss": 0.0731, "step": 2480 }, { "epoch": 0.4704111393357795, "grad_norm": 0.07986035943031311, "learning_rate": 0.0002, "loss": 0.0726, "step": 2500 }, { "epoch": 0.4741744284504657, "grad_norm": 0.07247906923294067, "learning_rate": 0.0002, "loss": 0.0763, "step": 2520 }, { "epoch": 0.47793771756515196, "grad_norm": 0.08120191842317581, "learning_rate": 0.0002, "loss": 0.0764, "step": 2540 }, { "epoch": 0.4817010066798382, "grad_norm": 0.09049087017774582, "learning_rate": 0.0002, "loss": 0.0799, "step": 2560 }, { "epoch": 0.48546429579452444, "grad_norm": 0.06859074532985687, "learning_rate": 0.0002, "loss": 0.0733, "step": 2580 }, { "epoch": 0.48922758490921064, "grad_norm": 0.07872766256332397, "learning_rate": 0.0002, "loss": 0.0724, "step": 2600 }, { "epoch": 0.4929908740238969, "grad_norm": 0.0663181021809578, "learning_rate": 0.0002, "loss": 0.0744, "step": 2620 }, { "epoch": 0.4967541631385831, "grad_norm": 0.07880811393260956, "learning_rate": 0.0002, "loss": 0.0759, "step": 2640 }, { "epoch": 0.5005174522532694, "grad_norm": 0.12468737363815308, "learning_rate": 0.0002, "loss": 0.0747, "step": 2660 }, { "epoch": 0.5042807413679556, "grad_norm": 0.08328507095575333, "learning_rate": 0.0002, "loss": 0.0754, "step": 2680 }, { "epoch": 0.5080440304826418, "grad_norm": 0.06665025651454926, "learning_rate": 0.0002, "loss": 0.0721, "step": 2700 }, { "epoch": 0.5118073195973281, "grad_norm": 0.07980209589004517, "learning_rate": 0.0002, "loss": 0.0733, "step": 2720 }, { "epoch": 0.5155706087120143, "grad_norm": 0.07951670140028, "learning_rate": 0.0002, "loss": 0.0792, "step": 2740 }, { "epoch": 0.5193338978267006, "grad_norm": 0.07515596598386765, "learning_rate": 0.0002, "loss": 0.0724, "step": 2760 }, { "epoch": 0.5230971869413867, "grad_norm": 0.08162270486354828, "learning_rate": 0.0002, "loss": 0.0714, "step": 2780 }, { "epoch": 0.526860476056073, "grad_norm": 0.08002326637506485, "learning_rate": 0.0002, "loss": 0.077, "step": 2800 }, { "epoch": 0.5306237651707593, "grad_norm": 0.07449716329574585, "learning_rate": 0.0002, "loss": 0.0707, "step": 2820 }, { "epoch": 0.5343870542854455, "grad_norm": 0.07990019023418427, "learning_rate": 0.0002, "loss": 0.071, "step": 2840 }, { "epoch": 0.5381503434001317, "grad_norm": 0.08538304269313812, "learning_rate": 0.0002, "loss": 0.0707, "step": 2860 }, { "epoch": 0.541913632514818, "grad_norm": 0.073255255818367, "learning_rate": 0.0002, "loss": 0.0699, "step": 2880 }, { "epoch": 0.5456769216295042, "grad_norm": 0.09101062268018723, "learning_rate": 0.0002, "loss": 0.0708, "step": 2900 }, { "epoch": 0.5494402107441905, "grad_norm": 0.0757337138056755, "learning_rate": 0.0002, "loss": 0.0702, "step": 2920 }, { "epoch": 0.5532034998588766, "grad_norm": 0.06553716212511063, "learning_rate": 0.0002, "loss": 0.0698, "step": 2940 }, { "epoch": 0.5569667889735629, "grad_norm": 0.08524072915315628, "learning_rate": 0.0002, "loss": 0.0738, "step": 2960 }, { "epoch": 0.5607300780882492, "grad_norm": 0.07361309975385666, "learning_rate": 0.0002, "loss": 0.0695, "step": 2980 }, { "epoch": 0.5644933672029354, "grad_norm": 0.07155182212591171, "learning_rate": 0.0002, "loss": 0.0667, "step": 3000 }, { "epoch": 0.5682566563176216, "grad_norm": 0.08088400214910507, "learning_rate": 0.0002, "loss": 0.0695, "step": 3020 }, { "epoch": 0.5720199454323078, "grad_norm": 0.08842818439006805, "learning_rate": 0.0002, "loss": 0.0656, "step": 3040 }, { "epoch": 0.5757832345469941, "grad_norm": 0.07135680317878723, "learning_rate": 0.0002, "loss": 0.0673, "step": 3060 }, { "epoch": 0.5795465236616804, "grad_norm": 0.08081484586000443, "learning_rate": 0.0002, "loss": 0.0709, "step": 3080 }, { "epoch": 0.5833098127763665, "grad_norm": 0.06999066472053528, "learning_rate": 0.0002, "loss": 0.0684, "step": 3100 }, { "epoch": 0.5870731018910528, "grad_norm": 0.06368447095155716, "learning_rate": 0.0002, "loss": 0.0691, "step": 3120 }, { "epoch": 0.590836391005739, "grad_norm": 0.08371565490961075, "learning_rate": 0.0002, "loss": 0.0641, "step": 3140 }, { "epoch": 0.5945996801204253, "grad_norm": 0.0675625279545784, "learning_rate": 0.0002, "loss": 0.069, "step": 3160 }, { "epoch": 0.5983629692351115, "grad_norm": 0.08121863007545471, "learning_rate": 0.0002, "loss": 0.065, "step": 3180 }, { "epoch": 0.6021262583497977, "grad_norm": 0.06689095497131348, "learning_rate": 0.0002, "loss": 0.0675, "step": 3200 }, { "epoch": 0.605889547464484, "grad_norm": 0.06634503602981567, "learning_rate": 0.0002, "loss": 0.064, "step": 3220 }, { "epoch": 0.6096528365791702, "grad_norm": 0.07062681019306183, "learning_rate": 0.0002, "loss": 0.0698, "step": 3240 }, { "epoch": 0.6134161256938564, "grad_norm": 0.07908321917057037, "learning_rate": 0.0002, "loss": 0.0674, "step": 3260 }, { "epoch": 0.6171794148085427, "grad_norm": 0.08132966607809067, "learning_rate": 0.0002, "loss": 0.0694, "step": 3280 }, { "epoch": 0.6209427039232289, "grad_norm": 0.081205353140831, "learning_rate": 0.0002, "loss": 0.0627, "step": 3300 }, { "epoch": 0.6247059930379152, "grad_norm": 0.07323930412530899, "learning_rate": 0.0002, "loss": 0.0645, "step": 3320 }, { "epoch": 0.6284692821526013, "grad_norm": 0.07979920506477356, "learning_rate": 0.0002, "loss": 0.0601, "step": 3340 }, { "epoch": 0.6322325712672876, "grad_norm": 0.06519381701946259, "learning_rate": 0.0002, "loss": 0.0677, "step": 3360 }, { "epoch": 0.6359958603819739, "grad_norm": 0.07861756533384323, "learning_rate": 0.0002, "loss": 0.0648, "step": 3380 }, { "epoch": 0.6397591494966601, "grad_norm": 0.09498826414346695, "learning_rate": 0.0002, "loss": 0.0644, "step": 3400 }, { "epoch": 0.6435224386113463, "grad_norm": 0.07069331407546997, "learning_rate": 0.0002, "loss": 0.0683, "step": 3420 }, { "epoch": 0.6472857277260325, "grad_norm": 0.07212232798337936, "learning_rate": 0.0002, "loss": 0.061, "step": 3440 }, { "epoch": 0.6510490168407188, "grad_norm": 0.06527985632419586, "learning_rate": 0.0002, "loss": 0.063, "step": 3460 }, { "epoch": 0.6548123059554051, "grad_norm": 0.09721993654966354, "learning_rate": 0.0002, "loss": 0.0699, "step": 3480 }, { "epoch": 0.6585755950700912, "grad_norm": 0.06381040811538696, "learning_rate": 0.0002, "loss": 0.0604, "step": 3500 }, { "epoch": 0.6623388841847775, "grad_norm": 0.07853078097105026, "learning_rate": 0.0002, "loss": 0.0649, "step": 3520 }, { "epoch": 0.6661021732994638, "grad_norm": 0.09212848544120789, "learning_rate": 0.0002, "loss": 0.0588, "step": 3540 }, { "epoch": 0.66986546241415, "grad_norm": 0.07097720354795456, "learning_rate": 0.0002, "loss": 0.0635, "step": 3560 }, { "epoch": 0.6736287515288362, "grad_norm": 0.08024760335683823, "learning_rate": 0.0002, "loss": 0.068, "step": 3580 }, { "epoch": 0.6773920406435224, "grad_norm": 0.07371170073747635, "learning_rate": 0.0002, "loss": 0.0652, "step": 3600 }, { "epoch": 0.6811553297582087, "grad_norm": 0.0894666537642479, "learning_rate": 0.0002, "loss": 0.0659, "step": 3620 }, { "epoch": 0.684918618872895, "grad_norm": 0.07533634454011917, "learning_rate": 0.0002, "loss": 0.0656, "step": 3640 }, { "epoch": 0.6886819079875811, "grad_norm": 0.09198255091905594, "learning_rate": 0.0002, "loss": 0.0645, "step": 3660 }, { "epoch": 0.6924451971022674, "grad_norm": 0.07213396579027176, "learning_rate": 0.0002, "loss": 0.0634, "step": 3680 }, { "epoch": 0.6962084862169536, "grad_norm": 0.06533551216125488, "learning_rate": 0.0002, "loss": 0.0666, "step": 3700 }, { "epoch": 0.6999717753316399, "grad_norm": 0.09535999596118927, "learning_rate": 0.0002, "loss": 0.0651, "step": 3720 }, { "epoch": 0.703735064446326, "grad_norm": 0.07631547003984451, "learning_rate": 0.0002, "loss": 0.0678, "step": 3740 }, { "epoch": 0.7074983535610123, "grad_norm": 0.09607011079788208, "learning_rate": 0.0002, "loss": 0.0681, "step": 3760 }, { "epoch": 0.7112616426756986, "grad_norm": 0.07209835946559906, "learning_rate": 0.0002, "loss": 0.068, "step": 3780 }, { "epoch": 0.7150249317903848, "grad_norm": 0.09137856960296631, "learning_rate": 0.0002, "loss": 0.0598, "step": 3800 }, { "epoch": 0.718788220905071, "grad_norm": 0.09964293241500854, "learning_rate": 0.0002, "loss": 0.0656, "step": 3820 }, { "epoch": 0.7225515100197573, "grad_norm": 0.07041902840137482, "learning_rate": 0.0002, "loss": 0.0617, "step": 3840 }, { "epoch": 0.7263147991344435, "grad_norm": 0.05418103560805321, "learning_rate": 0.0002, "loss": 0.0686, "step": 3860 }, { "epoch": 0.7300780882491298, "grad_norm": 0.07180003076791763, "learning_rate": 0.0002, "loss": 0.0627, "step": 3880 }, { "epoch": 0.7338413773638159, "grad_norm": 0.07342009246349335, "learning_rate": 0.0002, "loss": 0.0651, "step": 3900 }, { "epoch": 0.7376046664785022, "grad_norm": 0.08151030540466309, "learning_rate": 0.0002, "loss": 0.0645, "step": 3920 }, { "epoch": 0.7413679555931885, "grad_norm": 0.09977607429027557, "learning_rate": 0.0002, "loss": 0.0606, "step": 3940 }, { "epoch": 0.7451312447078747, "grad_norm": 0.06991346180438995, "learning_rate": 0.0002, "loss": 0.064, "step": 3960 }, { "epoch": 0.7488945338225609, "grad_norm": 0.06797617673873901, "learning_rate": 0.0002, "loss": 0.0595, "step": 3980 }, { "epoch": 0.7526578229372471, "grad_norm": 0.06056559830904007, "learning_rate": 0.0002, "loss": 0.0595, "step": 4000 }, { "epoch": 0.7564211120519334, "grad_norm": 0.06775388866662979, "learning_rate": 0.0002, "loss": 0.0624, "step": 4020 }, { "epoch": 0.7601844011666197, "grad_norm": 0.08091945201158524, "learning_rate": 0.0002, "loss": 0.0657, "step": 4040 }, { "epoch": 0.7639476902813058, "grad_norm": 0.1335289478302002, "learning_rate": 0.0002, "loss": 0.061, "step": 4060 }, { "epoch": 0.7677109793959921, "grad_norm": 0.07901336252689362, "learning_rate": 0.0002, "loss": 0.0635, "step": 4080 }, { "epoch": 0.7714742685106784, "grad_norm": 0.07874254137277603, "learning_rate": 0.0002, "loss": 0.066, "step": 4100 }, { "epoch": 0.7752375576253646, "grad_norm": 0.07547064125537872, "learning_rate": 0.0002, "loss": 0.061, "step": 4120 }, { "epoch": 0.7790008467400508, "grad_norm": 0.06392517685890198, "learning_rate": 0.0002, "loss": 0.0595, "step": 4140 }, { "epoch": 0.782764135854737, "grad_norm": 0.08115773648023605, "learning_rate": 0.0002, "loss": 0.0633, "step": 4160 }, { "epoch": 0.7865274249694233, "grad_norm": 0.07906658202409744, "learning_rate": 0.0002, "loss": 0.06, "step": 4180 }, { "epoch": 0.7902907140841096, "grad_norm": 0.8131846189498901, "learning_rate": 0.0002, "loss": 0.0642, "step": 4200 }, { "epoch": 0.7940540031987957, "grad_norm": 0.08816728740930557, "learning_rate": 0.0002, "loss": 0.0683, "step": 4220 }, { "epoch": 0.797817292313482, "grad_norm": 0.08981813490390778, "learning_rate": 0.0002, "loss": 0.0654, "step": 4240 }, { "epoch": 0.8015805814281682, "grad_norm": 0.08015542477369308, "learning_rate": 0.0002, "loss": 0.0623, "step": 4260 }, { "epoch": 0.8053438705428545, "grad_norm": 0.08604927361011505, "learning_rate": 0.0002, "loss": 0.0616, "step": 4280 }, { "epoch": 0.8091071596575407, "grad_norm": 0.08916622400283813, "learning_rate": 0.0002, "loss": 0.0651, "step": 4300 }, { "epoch": 0.8128704487722269, "grad_norm": 0.06978412717580795, "learning_rate": 0.0002, "loss": 0.0577, "step": 4320 }, { "epoch": 0.8166337378869132, "grad_norm": 0.089443139731884, "learning_rate": 0.0002, "loss": 0.0651, "step": 4340 }, { "epoch": 0.8203970270015994, "grad_norm": 0.08108926564455032, "learning_rate": 0.0002, "loss": 0.0584, "step": 4360 }, { "epoch": 0.8241603161162856, "grad_norm": 0.0913953110575676, "learning_rate": 0.0002, "loss": 0.0611, "step": 4380 }, { "epoch": 0.8279236052309719, "grad_norm": 0.07268033176660538, "learning_rate": 0.0002, "loss": 0.0601, "step": 4400 }, { "epoch": 0.8316868943456581, "grad_norm": 0.07602666318416595, "learning_rate": 0.0002, "loss": 0.0642, "step": 4420 }, { "epoch": 0.8354501834603444, "grad_norm": 0.06899157911539078, "learning_rate": 0.0002, "loss": 0.0603, "step": 4440 }, { "epoch": 0.8392134725750305, "grad_norm": 0.07310888916254044, "learning_rate": 0.0002, "loss": 0.0614, "step": 4460 }, { "epoch": 0.8429767616897168, "grad_norm": 0.086586132645607, "learning_rate": 0.0002, "loss": 0.0636, "step": 4480 }, { "epoch": 0.8467400508044031, "grad_norm": 0.04938528314232826, "learning_rate": 0.0002, "loss": 0.0577, "step": 4500 }, { "epoch": 0.8505033399190893, "grad_norm": 0.07192698121070862, "learning_rate": 0.0002, "loss": 0.0574, "step": 4520 }, { "epoch": 0.8542666290337755, "grad_norm": 0.06893257051706314, "learning_rate": 0.0002, "loss": 0.0604, "step": 4540 }, { "epoch": 0.8580299181484617, "grad_norm": 0.06691521406173706, "learning_rate": 0.0002, "loss": 0.0649, "step": 4560 }, { "epoch": 0.861793207263148, "grad_norm": 0.09420821070671082, "learning_rate": 0.0002, "loss": 0.0598, "step": 4580 }, { "epoch": 0.8655564963778343, "grad_norm": 0.07227174192667007, "learning_rate": 0.0002, "loss": 0.062, "step": 4600 }, { "epoch": 0.8693197854925204, "grad_norm": 0.07998435199260712, "learning_rate": 0.0002, "loss": 0.0585, "step": 4620 }, { "epoch": 0.8730830746072067, "grad_norm": 0.07850491255521774, "learning_rate": 0.0002, "loss": 0.0586, "step": 4640 }, { "epoch": 0.876846363721893, "grad_norm": 0.09450593590736389, "learning_rate": 0.0002, "loss": 0.0625, "step": 4660 }, { "epoch": 0.8806096528365792, "grad_norm": 0.08073689788579941, "learning_rate": 0.0002, "loss": 0.062, "step": 4680 }, { "epoch": 0.8843729419512654, "grad_norm": 0.06979519873857498, "learning_rate": 0.0002, "loss": 0.0584, "step": 4700 }, { "epoch": 0.8881362310659516, "grad_norm": 0.06407754868268967, "learning_rate": 0.0002, "loss": 0.0626, "step": 4720 }, { "epoch": 0.8918995201806379, "grad_norm": 0.2963426411151886, "learning_rate": 0.0002, "loss": 0.0625, "step": 4740 }, { "epoch": 0.8956628092953242, "grad_norm": 0.08902490139007568, "learning_rate": 0.0002, "loss": 0.0621, "step": 4760 }, { "epoch": 0.8994260984100103, "grad_norm": 0.08447249978780746, "learning_rate": 0.0002, "loss": 0.0638, "step": 4780 }, { "epoch": 0.9031893875246966, "grad_norm": 0.07621373981237411, "learning_rate": 0.0002, "loss": 0.0618, "step": 4800 }, { "epoch": 0.9069526766393828, "grad_norm": 0.07609863579273224, "learning_rate": 0.0002, "loss": 0.0627, "step": 4820 }, { "epoch": 0.9107159657540691, "grad_norm": 0.0785209983587265, "learning_rate": 0.0002, "loss": 0.0606, "step": 4840 }, { "epoch": 0.9144792548687553, "grad_norm": 0.08515089750289917, "learning_rate": 0.0002, "loss": 0.0577, "step": 4860 }, { "epoch": 0.9182425439834415, "grad_norm": 0.0810500979423523, "learning_rate": 0.0002, "loss": 0.0578, "step": 4880 }, { "epoch": 0.9220058330981278, "grad_norm": 0.08074364811182022, "learning_rate": 0.0002, "loss": 0.0556, "step": 4900 }, { "epoch": 0.925769122212814, "grad_norm": 0.06328209489583969, "learning_rate": 0.0002, "loss": 0.0555, "step": 4920 }, { "epoch": 0.9295324113275002, "grad_norm": 0.06556906551122665, "learning_rate": 0.0002, "loss": 0.056, "step": 4940 }, { "epoch": 0.9332957004421865, "grad_norm": 0.0747644379734993, "learning_rate": 0.0002, "loss": 0.0608, "step": 4960 }, { "epoch": 0.9370589895568727, "grad_norm": 0.07274675369262695, "learning_rate": 0.0002, "loss": 0.057, "step": 4980 }, { "epoch": 0.940822278671559, "grad_norm": 0.061261050403118134, "learning_rate": 0.0002, "loss": 0.0587, "step": 5000 }, { "epoch": 0.9445855677862451, "grad_norm": 0.08277834206819534, "learning_rate": 0.0002, "loss": 0.0608, "step": 5020 }, { "epoch": 0.9483488569009314, "grad_norm": 0.08663126826286316, "learning_rate": 0.0002, "loss": 0.0595, "step": 5040 }, { "epoch": 0.9521121460156177, "grad_norm": 0.06938447803258896, "learning_rate": 0.0002, "loss": 0.0567, "step": 5060 }, { "epoch": 0.9558754351303039, "grad_norm": 0.07865908741950989, "learning_rate": 0.0002, "loss": 0.0588, "step": 5080 }, { "epoch": 0.9596387242449901, "grad_norm": 0.061116304248571396, "learning_rate": 0.0002, "loss": 0.0574, "step": 5100 }, { "epoch": 0.9634020133596763, "grad_norm": 0.05689473822712898, "learning_rate": 0.0002, "loss": 0.0533, "step": 5120 }, { "epoch": 0.9671653024743626, "grad_norm": 0.07368452847003937, "learning_rate": 0.0002, "loss": 0.0591, "step": 5140 }, { "epoch": 0.9709285915890489, "grad_norm": 0.06758634001016617, "learning_rate": 0.0002, "loss": 0.0572, "step": 5160 }, { "epoch": 0.974691880703735, "grad_norm": 0.07348816096782684, "learning_rate": 0.0002, "loss": 0.0577, "step": 5180 }, { "epoch": 0.9784551698184213, "grad_norm": 0.07613357156515121, "learning_rate": 0.0002, "loss": 0.0566, "step": 5200 }, { "epoch": 0.9822184589331076, "grad_norm": 0.062316812574863434, "learning_rate": 0.0002, "loss": 0.0621, "step": 5220 }, { "epoch": 0.9859817480477938, "grad_norm": 0.08600709587335587, "learning_rate": 0.0002, "loss": 0.0576, "step": 5240 }, { "epoch": 0.98974503716248, "grad_norm": 0.06687742471694946, "learning_rate": 0.0002, "loss": 0.0571, "step": 5260 }, { "epoch": 0.9935083262771662, "grad_norm": 0.07486743479967117, "learning_rate": 0.0002, "loss": 0.0594, "step": 5280 }, { "epoch": 0.9972716153918525, "grad_norm": 0.08373595029115677, "learning_rate": 0.0002, "loss": 0.0579, "step": 5300 }, { "epoch": 0.9999059177721329, "eval_loss": 0.20169740915298462, "eval_runtime": 422.4603, "eval_samples_per_second": 4.1, "eval_steps_per_second": 0.514, "step": 5314 } ], "logging_steps": 20, "max_steps": 14000, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 77, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.423221588013613e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }