| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0135203650498563, | |
| "eval_steps": 500, | |
| "global_step": 1500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.006760182524928173, | |
| "grad_norm": 30.429955125606906, | |
| "learning_rate": 2.0270270270270273e-07, | |
| "loss": 1.0305, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.013520365049856346, | |
| "grad_norm": 11.031569316357162, | |
| "learning_rate": 4.27927927927928e-07, | |
| "loss": 0.9433, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.020280547574784518, | |
| "grad_norm": 7.568247729610925, | |
| "learning_rate": 6.531531531531532e-07, | |
| "loss": 0.7961, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.027040730099712692, | |
| "grad_norm": 7.13506773134569, | |
| "learning_rate": 8.783783783783785e-07, | |
| "loss": 0.6703, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.03380091262464086, | |
| "grad_norm": 5.13118234249973, | |
| "learning_rate": 1.1036036036036037e-06, | |
| "loss": 0.5868, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.040561095149569036, | |
| "grad_norm": 5.539727298943678, | |
| "learning_rate": 1.328828828828829e-06, | |
| "loss": 0.5074, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.04732127767449721, | |
| "grad_norm": 5.305287069709931, | |
| "learning_rate": 1.5540540540540541e-06, | |
| "loss": 0.4983, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.054081460199425384, | |
| "grad_norm": 5.106611798846999, | |
| "learning_rate": 1.7792792792792792e-06, | |
| "loss": 0.497, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.06084164272435356, | |
| "grad_norm": 4.582177338515729, | |
| "learning_rate": 2.0045045045045045e-06, | |
| "loss": 0.4695, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.06760182524928172, | |
| "grad_norm": 4.739272582163485, | |
| "learning_rate": 2.22972972972973e-06, | |
| "loss": 0.4408, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.0743620077742099, | |
| "grad_norm": 4.322112804201578, | |
| "learning_rate": 2.454954954954955e-06, | |
| "loss": 0.4215, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.08112219029913807, | |
| "grad_norm": 4.480237396840778, | |
| "learning_rate": 2.6801801801801803e-06, | |
| "loss": 0.4222, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.08788237282406625, | |
| "grad_norm": 4.588802640700157, | |
| "learning_rate": 2.9054054054054054e-06, | |
| "loss": 0.4297, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.09464255534899442, | |
| "grad_norm": 3.634384371273362, | |
| "learning_rate": 3.130630630630631e-06, | |
| "loss": 0.4132, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.1014027378739226, | |
| "grad_norm": 4.029617898426818, | |
| "learning_rate": 3.3558558558558565e-06, | |
| "loss": 0.4123, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.10816292039885077, | |
| "grad_norm": 3.6780967071398005, | |
| "learning_rate": 3.5810810810810816e-06, | |
| "loss": 0.4022, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.11492310292377894, | |
| "grad_norm": 3.8355903508595155, | |
| "learning_rate": 3.8063063063063067e-06, | |
| "loss": 0.417, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.12168328544870712, | |
| "grad_norm": 3.8796259132382604, | |
| "learning_rate": 4.031531531531531e-06, | |
| "loss": 0.4137, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.1284434679736353, | |
| "grad_norm": 3.513646073103006, | |
| "learning_rate": 4.256756756756757e-06, | |
| "loss": 0.3838, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.13520365049856345, | |
| "grad_norm": 3.280829869725974, | |
| "learning_rate": 4.4819819819819824e-06, | |
| "loss": 0.3933, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.14196383302349164, | |
| "grad_norm": 3.8684728373816877, | |
| "learning_rate": 4.707207207207208e-06, | |
| "loss": 0.418, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.1487240155484198, | |
| "grad_norm": 3.430012571317074, | |
| "learning_rate": 4.932432432432433e-06, | |
| "loss": 0.4068, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.15548419807334798, | |
| "grad_norm": 3.4236933785478594, | |
| "learning_rate": 5.157657657657657e-06, | |
| "loss": 0.4122, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.16224438059827614, | |
| "grad_norm": 3.337506294944149, | |
| "learning_rate": 5.382882882882884e-06, | |
| "loss": 0.4195, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.16900456312320433, | |
| "grad_norm": 4.340322941603959, | |
| "learning_rate": 5.608108108108109e-06, | |
| "loss": 0.3922, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.1757647456481325, | |
| "grad_norm": 3.079365176995162, | |
| "learning_rate": 5.833333333333334e-06, | |
| "loss": 0.3971, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.18252492817306068, | |
| "grad_norm": 3.1303889674723857, | |
| "learning_rate": 6.0585585585585595e-06, | |
| "loss": 0.4135, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.18928511069798884, | |
| "grad_norm": 3.2972741045149605, | |
| "learning_rate": 6.283783783783784e-06, | |
| "loss": 0.4021, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.19604529322291703, | |
| "grad_norm": 6.693630492235905, | |
| "learning_rate": 6.50900900900901e-06, | |
| "loss": 0.4159, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.2028054757478452, | |
| "grad_norm": 2.9729373757056456, | |
| "learning_rate": 6.734234234234235e-06, | |
| "loss": 0.4058, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.20956565827277338, | |
| "grad_norm": 2.95948849243511, | |
| "learning_rate": 6.95945945945946e-06, | |
| "loss": 0.4202, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.21632584079770154, | |
| "grad_norm": 2.8167543124734618, | |
| "learning_rate": 7.1846846846846855e-06, | |
| "loss": 0.4018, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.22308602332262972, | |
| "grad_norm": 2.8063934885840314, | |
| "learning_rate": 7.40990990990991e-06, | |
| "loss": 0.3997, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.22984620584755788, | |
| "grad_norm": 3.1516244053144, | |
| "learning_rate": 7.635135135135135e-06, | |
| "loss": 0.4172, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.23660638837248607, | |
| "grad_norm": 2.6885724493695715, | |
| "learning_rate": 7.860360360360361e-06, | |
| "loss": 0.3972, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.24336657089741423, | |
| "grad_norm": 2.6078839515038528, | |
| "learning_rate": 8.085585585585586e-06, | |
| "loss": 0.4076, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.2501267534223424, | |
| "grad_norm": 2.5241273124658434, | |
| "learning_rate": 8.31081081081081e-06, | |
| "loss": 0.4114, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.2568869359472706, | |
| "grad_norm": 2.811459731213271, | |
| "learning_rate": 8.536036036036037e-06, | |
| "loss": 0.397, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.26364711847219874, | |
| "grad_norm": 3.049745389493281, | |
| "learning_rate": 8.761261261261262e-06, | |
| "loss": 0.4213, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.2704073009971269, | |
| "grad_norm": 2.7296156561054303, | |
| "learning_rate": 8.986486486486488e-06, | |
| "loss": 0.42, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.2771674835220551, | |
| "grad_norm": 2.8072799681272285, | |
| "learning_rate": 9.211711711711713e-06, | |
| "loss": 0.4115, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.2839276660469833, | |
| "grad_norm": 2.45386826986572, | |
| "learning_rate": 9.436936936936937e-06, | |
| "loss": 0.4276, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.29068784857191143, | |
| "grad_norm": 2.6328350388772703, | |
| "learning_rate": 9.662162162162164e-06, | |
| "loss": 0.4195, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.2974480310968396, | |
| "grad_norm": 2.5446377303913392, | |
| "learning_rate": 9.887387387387388e-06, | |
| "loss": 0.4181, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.3042082136217678, | |
| "grad_norm": 2.5969527355663975, | |
| "learning_rate": 9.999961369685454e-06, | |
| "loss": 0.4128, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.31096839614669597, | |
| "grad_norm": 2.771780495890244, | |
| "learning_rate": 9.999652330750595e-06, | |
| "loss": 0.4055, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.31772857867162413, | |
| "grad_norm": 2.4553977895620336, | |
| "learning_rate": 9.99903427198204e-06, | |
| "loss": 0.4012, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.3244887611965523, | |
| "grad_norm": 2.4098745339341456, | |
| "learning_rate": 9.998107231580925e-06, | |
| "loss": 0.3968, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.3312489437214805, | |
| "grad_norm": 2.545785489303096, | |
| "learning_rate": 9.99687126684601e-06, | |
| "loss": 0.4192, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.33800912624640866, | |
| "grad_norm": 2.6063200322372575, | |
| "learning_rate": 9.995326454170132e-06, | |
| "loss": 0.4109, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.33800912624640866, | |
| "eval_loss": 0.41546937823295593, | |
| "eval_runtime": 1323.9234, | |
| "eval_samples_per_second": 3.973, | |
| "eval_steps_per_second": 0.497, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.3447693087713368, | |
| "grad_norm": 2.5250471936075582, | |
| "learning_rate": 9.993472889035478e-06, | |
| "loss": 0.4122, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.351529491296265, | |
| "grad_norm": 2.226305697507818, | |
| "learning_rate": 9.991310686007694e-06, | |
| "loss": 0.4068, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.3582896738211932, | |
| "grad_norm": 2.305497784945858, | |
| "learning_rate": 9.988839978728798e-06, | |
| "loss": 0.3922, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.36504985634612136, | |
| "grad_norm": 2.2810190229102267, | |
| "learning_rate": 9.986060919908917e-06, | |
| "loss": 0.4038, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.3718100388710495, | |
| "grad_norm": 2.0830775327743534, | |
| "learning_rate": 9.982973681316854e-06, | |
| "loss": 0.389, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.3785702213959777, | |
| "grad_norm": 2.412349310792309, | |
| "learning_rate": 9.97957845376947e-06, | |
| "loss": 0.4, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.38533040392090584, | |
| "grad_norm": 2.334930606027013, | |
| "learning_rate": 9.975875447119884e-06, | |
| "loss": 0.3925, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.39209058644583406, | |
| "grad_norm": 2.373633114437491, | |
| "learning_rate": 9.971864890244514e-06, | |
| "loss": 0.4038, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.3988507689707622, | |
| "grad_norm": 2.309096972224613, | |
| "learning_rate": 9.967547031028917e-06, | |
| "loss": 0.4008, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.4056109514956904, | |
| "grad_norm": 2.140294095106148, | |
| "learning_rate": 9.962922136352482e-06, | |
| "loss": 0.4004, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.41237113402061853, | |
| "grad_norm": 2.2849655710928287, | |
| "learning_rate": 9.957990492071917e-06, | |
| "loss": 0.4057, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.41913131654554675, | |
| "grad_norm": 2.092617163224422, | |
| "learning_rate": 9.9527524030036e-06, | |
| "loss": 0.4016, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.4258914990704749, | |
| "grad_norm": 2.786668476771917, | |
| "learning_rate": 9.947208192904722e-06, | |
| "loss": 0.3788, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.43265168159540307, | |
| "grad_norm": 2.1368728170135873, | |
| "learning_rate": 9.941358204453294e-06, | |
| "loss": 0.4022, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.43941186412033123, | |
| "grad_norm": 2.03479655017632, | |
| "learning_rate": 9.935202799226941e-06, | |
| "loss": 0.3795, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.44617204664525945, | |
| "grad_norm": 2.1100436195870964, | |
| "learning_rate": 9.928742357680586e-06, | |
| "loss": 0.3841, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.4529322291701876, | |
| "grad_norm": 2.1376422874279952, | |
| "learning_rate": 9.9219772791229e-06, | |
| "loss": 0.3809, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.45969241169511577, | |
| "grad_norm": 2.2131150187295865, | |
| "learning_rate": 9.914907981691656e-06, | |
| "loss": 0.3934, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.4664525942200439, | |
| "grad_norm": 2.5997656656639023, | |
| "learning_rate": 9.907534902327855e-06, | |
| "loss": 0.3975, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.47321277674497214, | |
| "grad_norm": 1.9856693804357932, | |
| "learning_rate": 9.899858496748738e-06, | |
| "loss": 0.38, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.4799729592699003, | |
| "grad_norm": 2.217516603205192, | |
| "learning_rate": 9.891879239419609e-06, | |
| "loss": 0.3789, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.48673314179482846, | |
| "grad_norm": 2.0702819723867427, | |
| "learning_rate": 9.883597623524518e-06, | |
| "loss": 0.3954, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.4934933243197566, | |
| "grad_norm": 2.2207526501051387, | |
| "learning_rate": 9.875014160935773e-06, | |
| "loss": 0.3816, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.5002535068446848, | |
| "grad_norm": 2.0307571044715487, | |
| "learning_rate": 9.866129382182295e-06, | |
| "loss": 0.3596, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.507013689369613, | |
| "grad_norm": 2.0484361446295756, | |
| "learning_rate": 9.856943836416844e-06, | |
| "loss": 0.3723, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.5137738718945412, | |
| "grad_norm": 2.188933520806837, | |
| "learning_rate": 9.847458091382057e-06, | |
| "loss": 0.3836, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.5205340544194693, | |
| "grad_norm": 2.2059318749133343, | |
| "learning_rate": 9.837672733375377e-06, | |
| "loss": 0.3764, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.5272942369443975, | |
| "grad_norm": 2.0738552350322657, | |
| "learning_rate": 9.827588367212797e-06, | |
| "loss": 0.3801, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.5340544194693256, | |
| "grad_norm": 2.049788797772694, | |
| "learning_rate": 9.81720561619149e-06, | |
| "loss": 0.3732, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.5408146019942538, | |
| "grad_norm": 2.607021547053881, | |
| "learning_rate": 9.806525122051276e-06, | |
| "loss": 0.3607, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.5475747845191821, | |
| "grad_norm": 2.0297130039478573, | |
| "learning_rate": 9.795547544934964e-06, | |
| "loss": 0.3756, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.5543349670441102, | |
| "grad_norm": 2.135186067946862, | |
| "learning_rate": 9.784273563347542e-06, | |
| "loss": 0.3597, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.5610951495690384, | |
| "grad_norm": 1.8995362331712584, | |
| "learning_rate": 9.772703874114246e-06, | |
| "loss": 0.3694, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.5678553320939665, | |
| "grad_norm": 2.0415002686577775, | |
| "learning_rate": 9.760839192337487e-06, | |
| "loss": 0.3799, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.5746155146188947, | |
| "grad_norm": 2.1273417637254366, | |
| "learning_rate": 9.74868025135266e-06, | |
| "loss": 0.3771, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.5813756971438229, | |
| "grad_norm": 2.1781086970923575, | |
| "learning_rate": 9.7362278026828e-06, | |
| "loss": 0.3817, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.588135879668751, | |
| "grad_norm": 2.1564198664014667, | |
| "learning_rate": 9.723482615992153e-06, | |
| "loss": 0.3755, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.5948960621936792, | |
| "grad_norm": 2.2895864922611993, | |
| "learning_rate": 9.710445479038585e-06, | |
| "loss": 0.379, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.6016562447186073, | |
| "grad_norm": 1.983081982245503, | |
| "learning_rate": 9.697117197624903e-06, | |
| "loss": 0.3648, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.6084164272435356, | |
| "grad_norm": 2.091041703194283, | |
| "learning_rate": 9.683498595549058e-06, | |
| "loss": 0.3537, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.6151766097684638, | |
| "grad_norm": 2.236195435477624, | |
| "learning_rate": 9.669590514553202e-06, | |
| "loss": 0.3706, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.6219367922933919, | |
| "grad_norm": 1.8500568790108252, | |
| "learning_rate": 9.65539381427169e-06, | |
| "loss": 0.3829, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.6286969748183201, | |
| "grad_norm": 1.9798512403639599, | |
| "learning_rate": 9.640909372177923e-06, | |
| "loss": 0.3583, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.6354571573432483, | |
| "grad_norm": 2.001295874575185, | |
| "learning_rate": 9.62613808353013e-06, | |
| "loss": 0.3846, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.6422173398681764, | |
| "grad_norm": 2.119068814752592, | |
| "learning_rate": 9.611080861316029e-06, | |
| "loss": 0.376, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.6489775223931046, | |
| "grad_norm": 1.874754216889785, | |
| "learning_rate": 9.595738636196392e-06, | |
| "loss": 0.3594, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.6557377049180327, | |
| "grad_norm": 1.9814730374903968, | |
| "learning_rate": 9.580112356447528e-06, | |
| "loss": 0.3763, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.662497887442961, | |
| "grad_norm": 1.6054885549770077, | |
| "learning_rate": 9.56420298790267e-06, | |
| "loss": 0.3539, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.6692580699678892, | |
| "grad_norm": 1.9862942162587933, | |
| "learning_rate": 9.548011513892274e-06, | |
| "loss": 0.3686, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.6760182524928173, | |
| "grad_norm": 1.697925964076465, | |
| "learning_rate": 9.531538935183252e-06, | |
| "loss": 0.3645, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.6760182524928173, | |
| "eval_loss": 0.36739876866340637, | |
| "eval_runtime": 1323.0599, | |
| "eval_samples_per_second": 3.976, | |
| "eval_steps_per_second": 0.497, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.6827784350177455, | |
| "grad_norm": 2.258646441282196, | |
| "learning_rate": 9.5147862699171e-06, | |
| "loss": 0.3686, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.6895386175426736, | |
| "grad_norm": 1.9875295083096067, | |
| "learning_rate": 9.497754553546992e-06, | |
| "loss": 0.3736, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.6962988000676018, | |
| "grad_norm": 1.9408896050189433, | |
| "learning_rate": 9.480444838773753e-06, | |
| "loss": 0.3542, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.70305898259253, | |
| "grad_norm": 2.517314352432714, | |
| "learning_rate": 9.462858195480814e-06, | |
| "loss": 0.3647, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.7098191651174581, | |
| "grad_norm": 2.045943843085583, | |
| "learning_rate": 9.444995710668074e-06, | |
| "loss": 0.3797, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.7165793476423864, | |
| "grad_norm": 1.9093836961309305, | |
| "learning_rate": 9.42685848838472e-06, | |
| "loss": 0.3413, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.7233395301673146, | |
| "grad_norm": 2.022174523080341, | |
| "learning_rate": 9.408447649660985e-06, | |
| "loss": 0.3676, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.7300997126922427, | |
| "grad_norm": 1.8275705704002196, | |
| "learning_rate": 9.38976433243886e-06, | |
| "loss": 0.3685, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.7368598952171709, | |
| "grad_norm": 1.7772669666798888, | |
| "learning_rate": 9.370809691501753e-06, | |
| "loss": 0.3708, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.743620077742099, | |
| "grad_norm": 1.9218973995164992, | |
| "learning_rate": 9.351584898403129e-06, | |
| "loss": 0.3546, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.7503802602670272, | |
| "grad_norm": 1.8585271029759325, | |
| "learning_rate": 9.332091141394082e-06, | |
| "loss": 0.351, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.7571404427919554, | |
| "grad_norm": 1.8921724238963678, | |
| "learning_rate": 9.312329625349903e-06, | |
| "loss": 0.3561, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.7639006253168835, | |
| "grad_norm": 2.699199676621496, | |
| "learning_rate": 9.292301571695603e-06, | |
| "loss": 0.3621, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.7706608078418117, | |
| "grad_norm": 1.7918758876099616, | |
| "learning_rate": 9.27200821833042e-06, | |
| "loss": 0.3481, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.77742099036674, | |
| "grad_norm": 1.6841732472542632, | |
| "learning_rate": 9.251450819551305e-06, | |
| "loss": 0.3518, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.7841811728916681, | |
| "grad_norm": 1.8263599680378415, | |
| "learning_rate": 9.2306306459754e-06, | |
| "loss": 0.3578, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.7909413554165963, | |
| "grad_norm": 1.9440999538041503, | |
| "learning_rate": 9.2095489844615e-06, | |
| "loss": 0.3561, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.7977015379415244, | |
| "grad_norm": 2.0375108827553725, | |
| "learning_rate": 9.188207138030518e-06, | |
| "loss": 0.3547, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.8044617204664526, | |
| "grad_norm": 1.8656977480472188, | |
| "learning_rate": 9.166606425784939e-06, | |
| "loss": 0.3353, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.8112219029913807, | |
| "grad_norm": 1.8340532352178343, | |
| "learning_rate": 9.144748182827305e-06, | |
| "loss": 0.3641, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.8179820855163089, | |
| "grad_norm": 1.8770982817901785, | |
| "learning_rate": 9.122633760177674e-06, | |
| "loss": 0.35, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.8247422680412371, | |
| "grad_norm": 1.7126903025921292, | |
| "learning_rate": 9.10026452469013e-06, | |
| "loss": 0.3483, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.8315024505661653, | |
| "grad_norm": 1.845094017702633, | |
| "learning_rate": 9.077641858968302e-06, | |
| "loss": 0.3463, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.8382626330910935, | |
| "grad_norm": 1.6919845905215602, | |
| "learning_rate": 9.054767161279891e-06, | |
| "loss": 0.3493, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.8450228156160217, | |
| "grad_norm": 41.7754958887055, | |
| "learning_rate": 9.031641845470265e-06, | |
| "loss": 0.3691, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.8517829981409498, | |
| "grad_norm": 1.6803994335239858, | |
| "learning_rate": 9.008267340875062e-06, | |
| "loss": 0.3404, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.858543180665878, | |
| "grad_norm": 1.6747949593133225, | |
| "learning_rate": 8.984645092231839e-06, | |
| "loss": 0.337, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.8653033631908061, | |
| "grad_norm": 1.9888542335284893, | |
| "learning_rate": 8.960776559590794e-06, | |
| "loss": 0.3531, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.8720635457157343, | |
| "grad_norm": 1.8242235825875925, | |
| "learning_rate": 8.936663218224504e-06, | |
| "loss": 0.3468, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.8788237282406625, | |
| "grad_norm": 1.8091464965495379, | |
| "learning_rate": 8.91230655853675e-06, | |
| "loss": 0.3451, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.8855839107655906, | |
| "grad_norm": 1.6077172216059503, | |
| "learning_rate": 8.887708085970395e-06, | |
| "loss": 0.3606, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.8923440932905189, | |
| "grad_norm": 1.863072270952894, | |
| "learning_rate": 8.862869320914342e-06, | |
| "loss": 0.3362, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.899104275815447, | |
| "grad_norm": 1.6689485544997376, | |
| "learning_rate": 8.837791798609548e-06, | |
| "loss": 0.3556, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.9058644583403752, | |
| "grad_norm": 1.7200717303760475, | |
| "learning_rate": 8.812477069054145e-06, | |
| "loss": 0.3478, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.9126246408653034, | |
| "grad_norm": 2.137634965409493, | |
| "learning_rate": 8.786926696907635e-06, | |
| "loss": 0.3367, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.9193848233902315, | |
| "grad_norm": 1.8766558650581284, | |
| "learning_rate": 8.761142261394176e-06, | |
| "loss": 0.3384, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.9261450059151597, | |
| "grad_norm": 1.79644209869537, | |
| "learning_rate": 8.735125356204982e-06, | |
| "loss": 0.3409, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.9329051884400879, | |
| "grad_norm": 1.878014748264182, | |
| "learning_rate": 8.708877589399805e-06, | |
| "loss": 0.3571, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.939665370965016, | |
| "grad_norm": 1.6990828286145954, | |
| "learning_rate": 8.682400583307562e-06, | |
| "loss": 0.3434, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.9464255534899443, | |
| "grad_norm": 1.7992025269775698, | |
| "learning_rate": 8.655695974426049e-06, | |
| "loss": 0.3469, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.9531857360148724, | |
| "grad_norm": 1.9318642460759288, | |
| "learning_rate": 8.628765413320793e-06, | |
| "loss": 0.3567, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.9599459185398006, | |
| "grad_norm": 1.7485812674142844, | |
| "learning_rate": 8.60161056452304e-06, | |
| "loss": 0.3295, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.9667061010647288, | |
| "grad_norm": 1.9314544861372662, | |
| "learning_rate": 8.574233106426866e-06, | |
| "loss": 0.331, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.9734662835896569, | |
| "grad_norm": 1.6993343654728899, | |
| "learning_rate": 8.546634731185446e-06, | |
| "loss": 0.353, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.9802264661145851, | |
| "grad_norm": 1.7502543394189403, | |
| "learning_rate": 8.518817144606451e-06, | |
| "loss": 0.3316, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.9869866486395132, | |
| "grad_norm": 1.9261600970610924, | |
| "learning_rate": 8.490782066046634e-06, | |
| "loss": 0.3628, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.9937468311644414, | |
| "grad_norm": 1.7416309439075797, | |
| "learning_rate": 8.462531228305546e-06, | |
| "loss": 0.3231, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 3.593217730508142, | |
| "learning_rate": 8.434066377518437e-06, | |
| "loss": 0.3394, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.0067601825249282, | |
| "grad_norm": 1.6132270094363823, | |
| "learning_rate": 8.405389273048334e-06, | |
| "loss": 0.2478, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.0135203650498563, | |
| "grad_norm": 1.6776338845512389, | |
| "learning_rate": 8.376501687377297e-06, | |
| "loss": 0.2299, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.0135203650498563, | |
| "eval_loss": 0.351112425327301, | |
| "eval_runtime": 1325.3556, | |
| "eval_samples_per_second": 3.969, | |
| "eval_steps_per_second": 0.496, | |
| "step": 1500 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 4440, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 576417156300800.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |