checkpoint-795 / trainer_state.json
tyzhu's picture
Uploading file checkpoint-795 to hftyzhu/tyzhu/checkpoint-795at time 2024-07-21 16:09:12
4bddc4f verified
raw
history blame
No virus
141 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9990771103736721,
"eval_steps": 50,
"global_step": 795,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0012567007677656253,
"grad_norm": 18.75091794627999,
"learning_rate": 2.5000000000000004e-07,
"loss": 0.7624,
"step": 1
},
{
"epoch": 0.0025134015355312506,
"grad_norm": 17.52654411272103,
"learning_rate": 5.000000000000001e-07,
"loss": 0.7516,
"step": 2
},
{
"epoch": 0.003770102303296876,
"grad_norm": 16.514255723157607,
"learning_rate": 7.5e-07,
"loss": 0.7638,
"step": 3
},
{
"epoch": 0.005026803071062501,
"grad_norm": 15.863246140843936,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.7628,
"step": 4
},
{
"epoch": 0.006283503838828127,
"grad_norm": 16.009485294186245,
"learning_rate": 1.25e-06,
"loss": 0.6841,
"step": 5
},
{
"epoch": 0.007540204606593752,
"grad_norm": 12.359798457900594,
"learning_rate": 1.5e-06,
"loss": 0.6439,
"step": 6
},
{
"epoch": 0.008796905374359378,
"grad_norm": 7.382396803359724,
"learning_rate": 1.75e-06,
"loss": 0.6556,
"step": 7
},
{
"epoch": 0.010053606142125002,
"grad_norm": 9.902433893820467,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.5683,
"step": 8
},
{
"epoch": 0.011310306909890627,
"grad_norm": 5.336530839540533,
"learning_rate": 2.25e-06,
"loss": 0.5813,
"step": 9
},
{
"epoch": 0.012567007677656254,
"grad_norm": 4.96442690704384,
"learning_rate": 2.5e-06,
"loss": 0.5555,
"step": 10
},
{
"epoch": 0.013823708445421879,
"grad_norm": 4.049027389723904,
"learning_rate": 2.7500000000000004e-06,
"loss": 0.5449,
"step": 11
},
{
"epoch": 0.015080409213187504,
"grad_norm": 2.9876344466981175,
"learning_rate": 3e-06,
"loss": 0.5739,
"step": 12
},
{
"epoch": 0.01633710998095313,
"grad_norm": 3.5573781991080535,
"learning_rate": 3.2500000000000002e-06,
"loss": 0.5489,
"step": 13
},
{
"epoch": 0.017593810748718755,
"grad_norm": 2.8413983357245605,
"learning_rate": 3.5e-06,
"loss": 0.5336,
"step": 14
},
{
"epoch": 0.01885051151648438,
"grad_norm": 3.025149469387222,
"learning_rate": 3.7500000000000005e-06,
"loss": 0.6036,
"step": 15
},
{
"epoch": 0.020107212284250005,
"grad_norm": 3.220040516331875,
"learning_rate": 4.000000000000001e-06,
"loss": 0.5046,
"step": 16
},
{
"epoch": 0.02136391305201563,
"grad_norm": 2.783851178663962,
"learning_rate": 4.25e-06,
"loss": 0.5365,
"step": 17
},
{
"epoch": 0.022620613819781255,
"grad_norm": 2.897929267846762,
"learning_rate": 4.5e-06,
"loss": 0.5198,
"step": 18
},
{
"epoch": 0.02387731458754688,
"grad_norm": 2.506735216769208,
"learning_rate": 4.75e-06,
"loss": 0.5094,
"step": 19
},
{
"epoch": 0.025134015355312508,
"grad_norm": 2.565147056156479,
"learning_rate": 5e-06,
"loss": 0.5352,
"step": 20
},
{
"epoch": 0.026390716123078133,
"grad_norm": 2.516902243933518,
"learning_rate": 5.2500000000000006e-06,
"loss": 0.5026,
"step": 21
},
{
"epoch": 0.027647416890843757,
"grad_norm": 2.7534715241704832,
"learning_rate": 5.500000000000001e-06,
"loss": 0.5186,
"step": 22
},
{
"epoch": 0.028904117658609382,
"grad_norm": 2.9905067284933606,
"learning_rate": 5.75e-06,
"loss": 0.5913,
"step": 23
},
{
"epoch": 0.030160818426375007,
"grad_norm": 2.3261883764698608,
"learning_rate": 6e-06,
"loss": 0.5275,
"step": 24
},
{
"epoch": 0.03141751919414063,
"grad_norm": 2.2175719666441704,
"learning_rate": 6.25e-06,
"loss": 0.4986,
"step": 25
},
{
"epoch": 0.03267421996190626,
"grad_norm": 2.499813549718401,
"learning_rate": 6.5000000000000004e-06,
"loss": 0.5014,
"step": 26
},
{
"epoch": 0.03393092072967188,
"grad_norm": 2.217010030707422,
"learning_rate": 6.750000000000001e-06,
"loss": 0.5205,
"step": 27
},
{
"epoch": 0.03518762149743751,
"grad_norm": 2.30544152290609,
"learning_rate": 7e-06,
"loss": 0.5144,
"step": 28
},
{
"epoch": 0.03644432226520313,
"grad_norm": 2.322128869370069,
"learning_rate": 7.25e-06,
"loss": 0.583,
"step": 29
},
{
"epoch": 0.03770102303296876,
"grad_norm": 2.326781634174622,
"learning_rate": 7.500000000000001e-06,
"loss": 0.4918,
"step": 30
},
{
"epoch": 0.03895772380073438,
"grad_norm": 2.4674314043738943,
"learning_rate": 7.75e-06,
"loss": 0.5194,
"step": 31
},
{
"epoch": 0.04021442456850001,
"grad_norm": 2.1812796298312707,
"learning_rate": 8.000000000000001e-06,
"loss": 0.5206,
"step": 32
},
{
"epoch": 0.04147112533626564,
"grad_norm": 2.1549349105457836,
"learning_rate": 8.25e-06,
"loss": 0.5916,
"step": 33
},
{
"epoch": 0.04272782610403126,
"grad_norm": 2.199454313573332,
"learning_rate": 8.5e-06,
"loss": 0.5276,
"step": 34
},
{
"epoch": 0.04398452687179689,
"grad_norm": 2.1700798684465537,
"learning_rate": 8.750000000000001e-06,
"loss": 0.5261,
"step": 35
},
{
"epoch": 0.04524122763956251,
"grad_norm": 2.0007451576632858,
"learning_rate": 9e-06,
"loss": 0.528,
"step": 36
},
{
"epoch": 0.04649792840732814,
"grad_norm": 2.1400416338070367,
"learning_rate": 9.250000000000001e-06,
"loss": 0.5444,
"step": 37
},
{
"epoch": 0.04775462917509376,
"grad_norm": 2.403488357912535,
"learning_rate": 9.5e-06,
"loss": 0.5155,
"step": 38
},
{
"epoch": 0.04901132994285939,
"grad_norm": 3.009219784165261,
"learning_rate": 9.75e-06,
"loss": 0.5153,
"step": 39
},
{
"epoch": 0.050268030710625015,
"grad_norm": 2.142435686354672,
"learning_rate": 1e-05,
"loss": 0.493,
"step": 40
},
{
"epoch": 0.05152473147839064,
"grad_norm": 2.239870156543961,
"learning_rate": 1.025e-05,
"loss": 0.5078,
"step": 41
},
{
"epoch": 0.052781432246156265,
"grad_norm": 2.6345987571093086,
"learning_rate": 1.0500000000000001e-05,
"loss": 0.4873,
"step": 42
},
{
"epoch": 0.05403813301392189,
"grad_norm": 1.9418390969701194,
"learning_rate": 1.075e-05,
"loss": 0.4965,
"step": 43
},
{
"epoch": 0.055294833781687515,
"grad_norm": 2.6081247589601193,
"learning_rate": 1.1000000000000001e-05,
"loss": 0.5281,
"step": 44
},
{
"epoch": 0.056551534549453136,
"grad_norm": 2.0666878694041433,
"learning_rate": 1.125e-05,
"loss": 0.5448,
"step": 45
},
{
"epoch": 0.057808235317218765,
"grad_norm": 2.404619928095954,
"learning_rate": 1.15e-05,
"loss": 0.4765,
"step": 46
},
{
"epoch": 0.059064936084984386,
"grad_norm": 2.0576801046013724,
"learning_rate": 1.1750000000000001e-05,
"loss": 0.5118,
"step": 47
},
{
"epoch": 0.060321636852750014,
"grad_norm": 2.3318752319695717,
"learning_rate": 1.2e-05,
"loss": 0.5178,
"step": 48
},
{
"epoch": 0.06157833762051564,
"grad_norm": 2.1469466151657293,
"learning_rate": 1.2250000000000001e-05,
"loss": 0.5127,
"step": 49
},
{
"epoch": 0.06283503838828126,
"grad_norm": 2.122786996441985,
"learning_rate": 1.25e-05,
"loss": 0.508,
"step": 50
},
{
"epoch": 0.06283503838828126,
"eval_loss": NaN,
"eval_runtime": 399.176,
"eval_samples_per_second": 20.83,
"eval_steps_per_second": 2.605,
"step": 50
},
{
"epoch": 0.06409173915604689,
"grad_norm": 2.1375097174172617,
"learning_rate": 1.275e-05,
"loss": 0.5397,
"step": 51
},
{
"epoch": 0.06534843992381252,
"grad_norm": 2.106065428482765,
"learning_rate": 1.3000000000000001e-05,
"loss": 0.4904,
"step": 52
},
{
"epoch": 0.06660514069157814,
"grad_norm": 2.0188566811543973,
"learning_rate": 1.325e-05,
"loss": 0.5276,
"step": 53
},
{
"epoch": 0.06786184145934376,
"grad_norm": 2.243686294713883,
"learning_rate": 1.3500000000000001e-05,
"loss": 0.604,
"step": 54
},
{
"epoch": 0.06911854222710939,
"grad_norm": 2.195249183257568,
"learning_rate": 1.375e-05,
"loss": 0.4981,
"step": 55
},
{
"epoch": 0.07037524299487502,
"grad_norm": 2.192582683330822,
"learning_rate": 1.4e-05,
"loss": 0.5028,
"step": 56
},
{
"epoch": 0.07163194376264065,
"grad_norm": 2.0627327989918185,
"learning_rate": 1.425e-05,
"loss": 0.5125,
"step": 57
},
{
"epoch": 0.07288864453040626,
"grad_norm": 1.872712329188738,
"learning_rate": 1.45e-05,
"loss": 0.508,
"step": 58
},
{
"epoch": 0.07414534529817189,
"grad_norm": 2.2251336046889008,
"learning_rate": 1.4750000000000003e-05,
"loss": 0.5025,
"step": 59
},
{
"epoch": 0.07540204606593752,
"grad_norm": 1.974200197833877,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.5279,
"step": 60
},
{
"epoch": 0.07665874683370315,
"grad_norm": 1.938388866679753,
"learning_rate": 1.525e-05,
"loss": 0.5284,
"step": 61
},
{
"epoch": 0.07791544760146876,
"grad_norm": 1.9081681403522774,
"learning_rate": 1.55e-05,
"loss": 0.587,
"step": 62
},
{
"epoch": 0.07917214836923439,
"grad_norm": 1.86283059738371,
"learning_rate": 1.575e-05,
"loss": 0.4899,
"step": 63
},
{
"epoch": 0.08042884913700002,
"grad_norm": 1.947746623023955,
"learning_rate": 1.6000000000000003e-05,
"loss": 0.5527,
"step": 64
},
{
"epoch": 0.08168554990476565,
"grad_norm": 1.9186877361719825,
"learning_rate": 1.6250000000000002e-05,
"loss": 0.4835,
"step": 65
},
{
"epoch": 0.08294225067253128,
"grad_norm": 2.044575571724766,
"learning_rate": 1.65e-05,
"loss": 0.5214,
"step": 66
},
{
"epoch": 0.08419895144029689,
"grad_norm": 2.1551282978146578,
"learning_rate": 1.675e-05,
"loss": 0.5338,
"step": 67
},
{
"epoch": 0.08545565220806252,
"grad_norm": 1.9391411629335615,
"learning_rate": 1.7e-05,
"loss": 0.5632,
"step": 68
},
{
"epoch": 0.08671235297582815,
"grad_norm": 1.8837181224290052,
"learning_rate": 1.7250000000000003e-05,
"loss": 0.6073,
"step": 69
},
{
"epoch": 0.08796905374359378,
"grad_norm": 2.234034052091539,
"learning_rate": 1.7500000000000002e-05,
"loss": 0.5692,
"step": 70
},
{
"epoch": 0.0892257545113594,
"grad_norm": 1.8949651791435709,
"learning_rate": 1.775e-05,
"loss": 0.5247,
"step": 71
},
{
"epoch": 0.09048245527912502,
"grad_norm": 1.7148270660492497,
"learning_rate": 1.8e-05,
"loss": 0.5223,
"step": 72
},
{
"epoch": 0.09173915604689065,
"grad_norm": 1.9445592195454198,
"learning_rate": 1.825e-05,
"loss": 0.5089,
"step": 73
},
{
"epoch": 0.09299585681465627,
"grad_norm": 1.8265253877418266,
"learning_rate": 1.8500000000000002e-05,
"loss": 0.5236,
"step": 74
},
{
"epoch": 0.0942525575824219,
"grad_norm": 1.8233199917702092,
"learning_rate": 1.8750000000000002e-05,
"loss": 0.5649,
"step": 75
},
{
"epoch": 0.09550925835018752,
"grad_norm": 2.2269645461783405,
"learning_rate": 1.9e-05,
"loss": 0.5303,
"step": 76
},
{
"epoch": 0.09676595911795315,
"grad_norm": 1.8067805055645487,
"learning_rate": 1.925e-05,
"loss": 0.5763,
"step": 77
},
{
"epoch": 0.09802265988571877,
"grad_norm": 2.1573534367398364,
"learning_rate": 1.95e-05,
"loss": 0.5314,
"step": 78
},
{
"epoch": 0.0992793606534844,
"grad_norm": 2.0382764916384546,
"learning_rate": 1.9750000000000002e-05,
"loss": 0.566,
"step": 79
},
{
"epoch": 0.10053606142125003,
"grad_norm": 1.8481169886281565,
"learning_rate": 2e-05,
"loss": 0.519,
"step": 80
},
{
"epoch": 0.10179276218901565,
"grad_norm": 2.0250286194462577,
"learning_rate": 1.9999903471186634e-05,
"loss": 0.5109,
"step": 81
},
{
"epoch": 0.10304946295678127,
"grad_norm": 1.9865218436602592,
"learning_rate": 1.9999613886610097e-05,
"loss": 0.5131,
"step": 82
},
{
"epoch": 0.1043061637245469,
"grad_norm": 2.146087026568179,
"learning_rate": 1.9999131251861037e-05,
"loss": 0.5479,
"step": 83
},
{
"epoch": 0.10556286449231253,
"grad_norm": 1.883855000830492,
"learning_rate": 1.999845557625709e-05,
"loss": 0.5515,
"step": 84
},
{
"epoch": 0.10681956526007814,
"grad_norm": 1.9686030378439734,
"learning_rate": 1.9997586872842683e-05,
"loss": 0.527,
"step": 85
},
{
"epoch": 0.10807626602784377,
"grad_norm": 2.36575606473121,
"learning_rate": 1.9996525158388804e-05,
"loss": 0.5402,
"step": 86
},
{
"epoch": 0.1093329667956094,
"grad_norm": 1.7575993387464999,
"learning_rate": 1.999527045339266e-05,
"loss": 0.5285,
"step": 87
},
{
"epoch": 0.11058966756337503,
"grad_norm": 2.385054737158702,
"learning_rate": 1.9993822782077282e-05,
"loss": 0.5904,
"step": 88
},
{
"epoch": 0.11184636833114066,
"grad_norm": 1.9263541895607634,
"learning_rate": 1.999218217239108e-05,
"loss": 0.5315,
"step": 89
},
{
"epoch": 0.11310306909890627,
"grad_norm": 1.985590164700294,
"learning_rate": 1.999034865600726e-05,
"loss": 0.5356,
"step": 90
},
{
"epoch": 0.1143597698666719,
"grad_norm": 1.8572569893256032,
"learning_rate": 1.998832226832327e-05,
"loss": 0.5492,
"step": 91
},
{
"epoch": 0.11561647063443753,
"grad_norm": 2.0556913179235305,
"learning_rate": 1.9986103048460056e-05,
"loss": 0.5274,
"step": 92
},
{
"epoch": 0.11687317140220316,
"grad_norm": 1.7389370827636832,
"learning_rate": 1.9983691039261358e-05,
"loss": 0.5498,
"step": 93
},
{
"epoch": 0.11812987216996877,
"grad_norm": 1.8846320386596898,
"learning_rate": 1.9981086287292853e-05,
"loss": 0.6494,
"step": 94
},
{
"epoch": 0.1193865729377344,
"grad_norm": 1.6471533089571895,
"learning_rate": 1.997828884284126e-05,
"loss": 0.5245,
"step": 95
},
{
"epoch": 0.12064327370550003,
"grad_norm": 1.993702988982152,
"learning_rate": 1.9975298759913382e-05,
"loss": 0.5261,
"step": 96
},
{
"epoch": 0.12189997447326566,
"grad_norm": 1.74703558113797,
"learning_rate": 1.9972116096235047e-05,
"loss": 0.5069,
"step": 97
},
{
"epoch": 0.12315667524103129,
"grad_norm": 1.8912653823700982,
"learning_rate": 1.996874091325001e-05,
"loss": 0.5378,
"step": 98
},
{
"epoch": 0.1244133760087969,
"grad_norm": 1.9270886670287957,
"learning_rate": 1.9965173276118747e-05,
"loss": 0.513,
"step": 99
},
{
"epoch": 0.12567007677656253,
"grad_norm": 1.9246733227266641,
"learning_rate": 1.9961413253717214e-05,
"loss": 0.5371,
"step": 100
},
{
"epoch": 0.12567007677656253,
"eval_loss": NaN,
"eval_runtime": 385.6745,
"eval_samples_per_second": 21.56,
"eval_steps_per_second": 2.697,
"step": 100
},
{
"epoch": 0.12692677754432816,
"grad_norm": 1.7630213216746193,
"learning_rate": 1.9957460918635513e-05,
"loss": 0.495,
"step": 101
},
{
"epoch": 0.12818347831209378,
"grad_norm": 1.676584473649907,
"learning_rate": 1.995331634717649e-05,
"loss": 0.5571,
"step": 102
},
{
"epoch": 0.1294401790798594,
"grad_norm": 1.7393851557554054,
"learning_rate": 1.9948979619354253e-05,
"loss": 0.5263,
"step": 103
},
{
"epoch": 0.13069687984762504,
"grad_norm": 1.5715249917558214,
"learning_rate": 1.994445081889264e-05,
"loss": 0.586,
"step": 104
},
{
"epoch": 0.13195358061539067,
"grad_norm": 1.7638743676149398,
"learning_rate": 1.99397300332236e-05,
"loss": 0.5371,
"step": 105
},
{
"epoch": 0.13321028138315627,
"grad_norm": 1.7333353036478043,
"learning_rate": 1.99348173534855e-05,
"loss": 0.5144,
"step": 106
},
{
"epoch": 0.1344669821509219,
"grad_norm": 1.6032418727100688,
"learning_rate": 1.9929712874521375e-05,
"loss": 0.5037,
"step": 107
},
{
"epoch": 0.13572368291868753,
"grad_norm": 1.6332164064062114,
"learning_rate": 1.9924416694877077e-05,
"loss": 0.5427,
"step": 108
},
{
"epoch": 0.13698038368645316,
"grad_norm": 1.509888795711472,
"learning_rate": 1.9918928916799395e-05,
"loss": 0.5748,
"step": 109
},
{
"epoch": 0.13823708445421878,
"grad_norm": 1.7081555607417591,
"learning_rate": 1.9913249646234072e-05,
"loss": 0.5201,
"step": 110
},
{
"epoch": 0.1394937852219844,
"grad_norm": 1.5103689076557207,
"learning_rate": 1.9907378992823755e-05,
"loss": 0.5313,
"step": 111
},
{
"epoch": 0.14075048598975004,
"grad_norm": 1.6174363391736,
"learning_rate": 1.990131706990589e-05,
"loss": 0.5414,
"step": 112
},
{
"epoch": 0.14200718675751567,
"grad_norm": 1.5406431804458636,
"learning_rate": 1.9895063994510512e-05,
"loss": 0.5183,
"step": 113
},
{
"epoch": 0.1432638875252813,
"grad_norm": 1.5953704972891425,
"learning_rate": 1.9888619887358024e-05,
"loss": 0.5036,
"step": 114
},
{
"epoch": 0.1445205882930469,
"grad_norm": 1.6126627995125455,
"learning_rate": 1.988198487285682e-05,
"loss": 0.5268,
"step": 115
},
{
"epoch": 0.14577728906081253,
"grad_norm": 1.5872606292238929,
"learning_rate": 1.9875159079100917e-05,
"loss": 0.5404,
"step": 116
},
{
"epoch": 0.14703398982857815,
"grad_norm": 1.4997817516498149,
"learning_rate": 1.9868142637867474e-05,
"loss": 0.4942,
"step": 117
},
{
"epoch": 0.14829069059634378,
"grad_norm": 1.5577686391641898,
"learning_rate": 1.9860935684614235e-05,
"loss": 0.6059,
"step": 118
},
{
"epoch": 0.1495473913641094,
"grad_norm": 1.562098718613041,
"learning_rate": 1.9853538358476933e-05,
"loss": 0.5772,
"step": 119
},
{
"epoch": 0.15080409213187504,
"grad_norm": 1.4588688366756248,
"learning_rate": 1.9845950802266584e-05,
"loss": 0.5652,
"step": 120
},
{
"epoch": 0.15206079289964067,
"grad_norm": 1.4789008945019497,
"learning_rate": 1.983817316246676e-05,
"loss": 0.5258,
"step": 121
},
{
"epoch": 0.1533174936674063,
"grad_norm": 1.6321791739327218,
"learning_rate": 1.9830205589230713e-05,
"loss": 0.5833,
"step": 122
},
{
"epoch": 0.15457419443517192,
"grad_norm": 1.4311215238919617,
"learning_rate": 1.9822048236378536e-05,
"loss": 0.5357,
"step": 123
},
{
"epoch": 0.15583089520293752,
"grad_norm": 1.4711622354911698,
"learning_rate": 1.9813701261394136e-05,
"loss": 0.5412,
"step": 124
},
{
"epoch": 0.15708759597070315,
"grad_norm": 1.561014461052019,
"learning_rate": 1.980516482542224e-05,
"loss": 0.5579,
"step": 125
},
{
"epoch": 0.15834429673846878,
"grad_norm": 1.5602653545497562,
"learning_rate": 1.9796439093265245e-05,
"loss": 0.5243,
"step": 126
},
{
"epoch": 0.1596009975062344,
"grad_norm": 1.533314694761247,
"learning_rate": 1.9787524233380076e-05,
"loss": 0.525,
"step": 127
},
{
"epoch": 0.16085769827400004,
"grad_norm": 1.557676346471582,
"learning_rate": 1.9778420417874894e-05,
"loss": 0.5242,
"step": 128
},
{
"epoch": 0.16211439904176567,
"grad_norm": 1.5308931674522483,
"learning_rate": 1.9769127822505805e-05,
"loss": 0.6101,
"step": 129
},
{
"epoch": 0.1633710998095313,
"grad_norm": 1.588409977024727,
"learning_rate": 1.9759646626673445e-05,
"loss": 0.5748,
"step": 130
},
{
"epoch": 0.16462780057729692,
"grad_norm": 1.5825263108270393,
"learning_rate": 1.9749977013419536e-05,
"loss": 0.5317,
"step": 131
},
{
"epoch": 0.16588450134506255,
"grad_norm": 1.3805626708260326,
"learning_rate": 1.9740119169423337e-05,
"loss": 0.5818,
"step": 132
},
{
"epoch": 0.16714120211282815,
"grad_norm": 18.459816671434435,
"learning_rate": 1.973007328499804e-05,
"loss": 0.5388,
"step": 133
},
{
"epoch": 0.16839790288059378,
"grad_norm": 1.8855793357632873,
"learning_rate": 1.9719839554087108e-05,
"loss": 0.623,
"step": 134
},
{
"epoch": 0.1696546036483594,
"grad_norm": 1.542977480320436,
"learning_rate": 1.9709418174260523e-05,
"loss": 0.5248,
"step": 135
},
{
"epoch": 0.17091130441612504,
"grad_norm": 1.6917497316960342,
"learning_rate": 1.9698809346710965e-05,
"loss": 0.5624,
"step": 136
},
{
"epoch": 0.17216800518389067,
"grad_norm": 1.6100252189793525,
"learning_rate": 1.9688013276249947e-05,
"loss": 0.553,
"step": 137
},
{
"epoch": 0.1734247059516563,
"grad_norm": 1.5825712812861108,
"learning_rate": 1.9677030171303842e-05,
"loss": 0.5645,
"step": 138
},
{
"epoch": 0.17468140671942192,
"grad_norm": 1.6638750857270868,
"learning_rate": 1.966586024390986e-05,
"loss": 0.5349,
"step": 139
},
{
"epoch": 0.17593810748718755,
"grad_norm": 1.5448265769685363,
"learning_rate": 1.9654503709711984e-05,
"loss": 0.534,
"step": 140
},
{
"epoch": 0.17719480825495318,
"grad_norm": 1.605020502262543,
"learning_rate": 1.964296078795675e-05,
"loss": 0.5553,
"step": 141
},
{
"epoch": 0.1784515090227188,
"grad_norm": 1.527407593624494,
"learning_rate": 1.9631231701489083e-05,
"loss": 0.5298,
"step": 142
},
{
"epoch": 0.1797082097904844,
"grad_norm": 1.5204305071770103,
"learning_rate": 1.9619316676747928e-05,
"loss": 0.5212,
"step": 143
},
{
"epoch": 0.18096491055825004,
"grad_norm": 1.5980803059820479,
"learning_rate": 1.9607215943761933e-05,
"loss": 0.5611,
"step": 144
},
{
"epoch": 0.18222161132601566,
"grad_norm": 2.3142581133230986,
"learning_rate": 1.9594929736144978e-05,
"loss": 0.536,
"step": 145
},
{
"epoch": 0.1834783120937813,
"grad_norm": 1.5920429437394876,
"learning_rate": 1.9582458291091664e-05,
"loss": 0.5363,
"step": 146
},
{
"epoch": 0.18473501286154692,
"grad_norm": 1.4799465577039153,
"learning_rate": 1.9569801849372757e-05,
"loss": 0.5093,
"step": 147
},
{
"epoch": 0.18599171362931255,
"grad_norm": 1.5040620797412507,
"learning_rate": 1.9556960655330512e-05,
"loss": 0.5755,
"step": 148
},
{
"epoch": 0.18724841439707818,
"grad_norm": 1.5753521873652412,
"learning_rate": 1.954393495687398e-05,
"loss": 0.5408,
"step": 149
},
{
"epoch": 0.1885051151648438,
"grad_norm": 1.436217689781534,
"learning_rate": 1.9530725005474195e-05,
"loss": 0.5908,
"step": 150
},
{
"epoch": 0.1885051151648438,
"eval_loss": NaN,
"eval_runtime": 389.6633,
"eval_samples_per_second": 21.339,
"eval_steps_per_second": 2.669,
"step": 150
},
{
"epoch": 0.18976181593260943,
"grad_norm": 1.4915607993777378,
"learning_rate": 1.9517331056159353e-05,
"loss": 0.5643,
"step": 151
},
{
"epoch": 0.19101851670037504,
"grad_norm": 1.3658845196609777,
"learning_rate": 1.9503753367509855e-05,
"loss": 0.5409,
"step": 152
},
{
"epoch": 0.19227521746814066,
"grad_norm": 1.6396983118066124,
"learning_rate": 1.9489992201653337e-05,
"loss": 0.5318,
"step": 153
},
{
"epoch": 0.1935319182359063,
"grad_norm": 1.4530656863997269,
"learning_rate": 1.9476047824259602e-05,
"loss": 0.5288,
"step": 154
},
{
"epoch": 0.19478861900367192,
"grad_norm": 1.8323995061487757,
"learning_rate": 1.946192050453549e-05,
"loss": 0.5207,
"step": 155
},
{
"epoch": 0.19604531977143755,
"grad_norm": 1.4963171570508917,
"learning_rate": 1.944761051521968e-05,
"loss": 0.5341,
"step": 156
},
{
"epoch": 0.19730202053920318,
"grad_norm": 1.4571814574677353,
"learning_rate": 1.9433118132577432e-05,
"loss": 0.5147,
"step": 157
},
{
"epoch": 0.1985587213069688,
"grad_norm": 1.46930107978953,
"learning_rate": 1.941844363639525e-05,
"loss": 0.5533,
"step": 158
},
{
"epoch": 0.19981542207473443,
"grad_norm": 1.4876593945156182,
"learning_rate": 1.9403587309975467e-05,
"loss": 0.5524,
"step": 159
},
{
"epoch": 0.20107212284250006,
"grad_norm": 1.5458825520839694,
"learning_rate": 1.93885494401308e-05,
"loss": 0.5147,
"step": 160
},
{
"epoch": 0.20232882361026566,
"grad_norm": 1.4541059075079896,
"learning_rate": 1.9373330317178797e-05,
"loss": 0.6548,
"step": 161
},
{
"epoch": 0.2035855243780313,
"grad_norm": 1.4441733330016775,
"learning_rate": 1.9357930234936228e-05,
"loss": 0.5307,
"step": 162
},
{
"epoch": 0.20484222514579692,
"grad_norm": 1.3924614428073763,
"learning_rate": 1.9342349490713427e-05,
"loss": 0.4934,
"step": 163
},
{
"epoch": 0.20609892591356255,
"grad_norm": 1.4163101626594428,
"learning_rate": 1.932658838530855e-05,
"loss": 0.5248,
"step": 164
},
{
"epoch": 0.20735562668132818,
"grad_norm": 1.4434173531044028,
"learning_rate": 1.9310647223001752e-05,
"loss": 0.5606,
"step": 165
},
{
"epoch": 0.2086123274490938,
"grad_norm": 1.4428551480982263,
"learning_rate": 1.929452631154933e-05,
"loss": 0.531,
"step": 166
},
{
"epoch": 0.20986902821685943,
"grad_norm": 1.423851316496012,
"learning_rate": 1.9278225962177776e-05,
"loss": 0.5223,
"step": 167
},
{
"epoch": 0.21112572898462506,
"grad_norm": 1.492507341069709,
"learning_rate": 1.9261746489577767e-05,
"loss": 0.6371,
"step": 168
},
{
"epoch": 0.2123824297523907,
"grad_norm": 1.448145384070868,
"learning_rate": 1.9245088211898086e-05,
"loss": 0.5101,
"step": 169
},
{
"epoch": 0.2136391305201563,
"grad_norm": 1.384046704220895,
"learning_rate": 1.9228251450739495e-05,
"loss": 0.5975,
"step": 170
},
{
"epoch": 0.21489583128792192,
"grad_norm": 1.3991515824529313,
"learning_rate": 1.92112365311485e-05,
"loss": 0.5913,
"step": 171
},
{
"epoch": 0.21615253205568755,
"grad_norm": 1.3857833501335572,
"learning_rate": 1.919404378161111e-05,
"loss": 0.5224,
"step": 172
},
{
"epoch": 0.21740923282345317,
"grad_norm": 1.3507684127877164,
"learning_rate": 1.9176673534046465e-05,
"loss": 0.5393,
"step": 173
},
{
"epoch": 0.2186659335912188,
"grad_norm": 1.4477184992508454,
"learning_rate": 1.9159126123800437e-05,
"loss": 0.5426,
"step": 174
},
{
"epoch": 0.21992263435898443,
"grad_norm": 1.2834792398929344,
"learning_rate": 1.9141401889639167e-05,
"loss": 0.5497,
"step": 175
},
{
"epoch": 0.22117933512675006,
"grad_norm": 1.3623159756756789,
"learning_rate": 1.9123501173742514e-05,
"loss": 0.5103,
"step": 176
},
{
"epoch": 0.2224360358945157,
"grad_norm": 1.4467854185777025,
"learning_rate": 1.910542432169745e-05,
"loss": 0.5348,
"step": 177
},
{
"epoch": 0.22369273666228132,
"grad_norm": 1.3284354351981553,
"learning_rate": 1.908717168249139e-05,
"loss": 0.528,
"step": 178
},
{
"epoch": 0.22494943743004692,
"grad_norm": 1.397962009852414,
"learning_rate": 1.9068743608505454e-05,
"loss": 0.5727,
"step": 179
},
{
"epoch": 0.22620613819781255,
"grad_norm": 1.427342226625829,
"learning_rate": 1.905014045550767e-05,
"loss": 0.5328,
"step": 180
},
{
"epoch": 0.22746283896557817,
"grad_norm": 1.287239450182925,
"learning_rate": 1.903136258264609e-05,
"loss": 0.5404,
"step": 181
},
{
"epoch": 0.2287195397333438,
"grad_norm": 1.3231711055343047,
"learning_rate": 1.9012410352441866e-05,
"loss": 0.5387,
"step": 182
},
{
"epoch": 0.22997624050110943,
"grad_norm": 1.380871461116662,
"learning_rate": 1.899328413078227e-05,
"loss": 0.5245,
"step": 183
},
{
"epoch": 0.23123294126887506,
"grad_norm": 1.3476601420913,
"learning_rate": 1.8973984286913584e-05,
"loss": 0.5639,
"step": 184
},
{
"epoch": 0.2324896420366407,
"grad_norm": 1.3713222215493124,
"learning_rate": 1.8954511193434024e-05,
"loss": 0.533,
"step": 185
},
{
"epoch": 0.23374634280440632,
"grad_norm": 1.301173886904826,
"learning_rate": 1.8934865226286507e-05,
"loss": 0.5302,
"step": 186
},
{
"epoch": 0.23500304357217194,
"grad_norm": 1.250844193542083,
"learning_rate": 1.891504676475141e-05,
"loss": 0.5678,
"step": 187
},
{
"epoch": 0.23625974433993754,
"grad_norm": 1.3624917616288885,
"learning_rate": 1.8895056191439252e-05,
"loss": 0.5478,
"step": 188
},
{
"epoch": 0.23751644510770317,
"grad_norm": 1.316550725145833,
"learning_rate": 1.8874893892283296e-05,
"loss": 0.5371,
"step": 189
},
{
"epoch": 0.2387731458754688,
"grad_norm": 1.4024386573064895,
"learning_rate": 1.8854560256532098e-05,
"loss": 0.5223,
"step": 190
},
{
"epoch": 0.24002984664323443,
"grad_norm": 1.2545561250193908,
"learning_rate": 1.8834055676742018e-05,
"loss": 0.5312,
"step": 191
},
{
"epoch": 0.24128654741100006,
"grad_norm": 1.3191673233181607,
"learning_rate": 1.8813380548769594e-05,
"loss": 0.5449,
"step": 192
},
{
"epoch": 0.24254324817876569,
"grad_norm": 1.351413784775738,
"learning_rate": 1.8792535271763944e-05,
"loss": 0.5338,
"step": 193
},
{
"epoch": 0.24379994894653131,
"grad_norm": 1.241988064092758,
"learning_rate": 1.8771520248159044e-05,
"loss": 0.537,
"step": 194
},
{
"epoch": 0.24505664971429694,
"grad_norm": 1.2555364071859036,
"learning_rate": 1.8750335883665948e-05,
"loss": 0.5831,
"step": 195
},
{
"epoch": 0.24631335048206257,
"grad_norm": 1.2648163083865465,
"learning_rate": 1.8728982587264965e-05,
"loss": 0.5553,
"step": 196
},
{
"epoch": 0.2475700512498282,
"grad_norm": 1.3009972836834804,
"learning_rate": 1.8707460771197773e-05,
"loss": 0.5213,
"step": 197
},
{
"epoch": 0.2488267520175938,
"grad_norm": 1.308271909393604,
"learning_rate": 1.8685770850959444e-05,
"loss": 0.5781,
"step": 198
},
{
"epoch": 0.25008345278535943,
"grad_norm": 1.2985339297627132,
"learning_rate": 1.8663913245290433e-05,
"loss": 0.5319,
"step": 199
},
{
"epoch": 0.25134015355312506,
"grad_norm": 1.2451217830797545,
"learning_rate": 1.8641888376168483e-05,
"loss": 0.5608,
"step": 200
},
{
"epoch": 0.25134015355312506,
"eval_loss": NaN,
"eval_runtime": 381.9917,
"eval_samples_per_second": 21.767,
"eval_steps_per_second": 2.723,
"step": 200
},
{
"epoch": 0.2525968543208907,
"grad_norm": 1.4207395274648773,
"learning_rate": 1.8619696668800494e-05,
"loss": 0.5194,
"step": 201
},
{
"epoch": 0.2538535550886563,
"grad_norm": 1.2969288642452035,
"learning_rate": 1.85973385516143e-05,
"loss": 0.5303,
"step": 202
},
{
"epoch": 0.25511025585642194,
"grad_norm": 1.244817385015466,
"learning_rate": 1.8574814456250406e-05,
"loss": 0.5519,
"step": 203
},
{
"epoch": 0.25636695662418757,
"grad_norm": 1.2786899670392922,
"learning_rate": 1.855212481755365e-05,
"loss": 0.5314,
"step": 204
},
{
"epoch": 0.2576236573919532,
"grad_norm": 1.6119075958517899,
"learning_rate": 1.852927007356481e-05,
"loss": 0.5153,
"step": 205
},
{
"epoch": 0.2588803581597188,
"grad_norm": 1.3027579835871468,
"learning_rate": 1.8506250665512156e-05,
"loss": 0.5317,
"step": 206
},
{
"epoch": 0.26013705892748445,
"grad_norm": 1.4097127031945536,
"learning_rate": 1.848306703780291e-05,
"loss": 0.5432,
"step": 207
},
{
"epoch": 0.2613937596952501,
"grad_norm": 1.2253893934005955,
"learning_rate": 1.8459719638014693e-05,
"loss": 0.5144,
"step": 208
},
{
"epoch": 0.2626504604630157,
"grad_norm": 1.3490881789862665,
"learning_rate": 1.843620891688686e-05,
"loss": 0.5227,
"step": 209
},
{
"epoch": 0.26390716123078134,
"grad_norm": 1.2692047638329742,
"learning_rate": 1.8412535328311813e-05,
"loss": 0.5172,
"step": 210
},
{
"epoch": 0.2651638619985469,
"grad_norm": 1.3001969775108575,
"learning_rate": 1.8388699329326237e-05,
"loss": 0.5013,
"step": 211
},
{
"epoch": 0.26642056276631254,
"grad_norm": 1.2460584486023816,
"learning_rate": 1.8364701380102267e-05,
"loss": 0.5097,
"step": 212
},
{
"epoch": 0.26767726353407817,
"grad_norm": 1.2575445422859304,
"learning_rate": 1.8340541943938623e-05,
"loss": 0.5253,
"step": 213
},
{
"epoch": 0.2689339643018438,
"grad_norm": 1.247709368946671,
"learning_rate": 1.831622148725164e-05,
"loss": 0.5195,
"step": 214
},
{
"epoch": 0.2701906650696094,
"grad_norm": 1.331986681061102,
"learning_rate": 1.8291740479566286e-05,
"loss": 0.517,
"step": 215
},
{
"epoch": 0.27144736583737505,
"grad_norm": 1.2798935292338272,
"learning_rate": 1.8267099393507083e-05,
"loss": 0.5774,
"step": 216
},
{
"epoch": 0.2727040666051407,
"grad_norm": 1.2305947859868307,
"learning_rate": 1.8242298704788988e-05,
"loss": 0.5285,
"step": 217
},
{
"epoch": 0.2739607673729063,
"grad_norm": 1.2890582922664517,
"learning_rate": 1.821733889220822e-05,
"loss": 0.5111,
"step": 218
},
{
"epoch": 0.27521746814067194,
"grad_norm": 1.2655404616013888,
"learning_rate": 1.819222043763299e-05,
"loss": 0.549,
"step": 219
},
{
"epoch": 0.27647416890843757,
"grad_norm": 1.3037267105348513,
"learning_rate": 1.816694382599422e-05,
"loss": 0.535,
"step": 220
},
{
"epoch": 0.2777308696762032,
"grad_norm": 1.2775883338383642,
"learning_rate": 1.814150954527618e-05,
"loss": 0.518,
"step": 221
},
{
"epoch": 0.2789875704439688,
"grad_norm": 1.3182350382629517,
"learning_rate": 1.8115918086507054e-05,
"loss": 0.5879,
"step": 222
},
{
"epoch": 0.28024427121173445,
"grad_norm": 1.2837663456694282,
"learning_rate": 1.8090169943749477e-05,
"loss": 0.5575,
"step": 223
},
{
"epoch": 0.2815009719795001,
"grad_norm": 1.3410051389307014,
"learning_rate": 1.8064265614090976e-05,
"loss": 0.5776,
"step": 224
},
{
"epoch": 0.2827576727472657,
"grad_norm": 1.3538997044655456,
"learning_rate": 1.8038205597634392e-05,
"loss": 0.5315,
"step": 225
},
{
"epoch": 0.28401437351503134,
"grad_norm": 1.2837659278037532,
"learning_rate": 1.801199039748822e-05,
"loss": 0.5341,
"step": 226
},
{
"epoch": 0.28527107428279697,
"grad_norm": 1.3959368719539762,
"learning_rate": 1.7985620519756897e-05,
"loss": 0.53,
"step": 227
},
{
"epoch": 0.2865277750505626,
"grad_norm": 1.277924618484306,
"learning_rate": 1.7959096473531016e-05,
"loss": 0.5131,
"step": 228
},
{
"epoch": 0.2877844758183282,
"grad_norm": 1.360578070388895,
"learning_rate": 1.7932418770877523e-05,
"loss": 0.5146,
"step": 229
},
{
"epoch": 0.2890411765860938,
"grad_norm": 1.2558461807056212,
"learning_rate": 1.7905587926829815e-05,
"loss": 0.567,
"step": 230
},
{
"epoch": 0.2902978773538594,
"grad_norm": 1.2185312710624743,
"learning_rate": 1.7878604459377795e-05,
"loss": 0.5259,
"step": 231
},
{
"epoch": 0.29155457812162505,
"grad_norm": 1.1707487670839811,
"learning_rate": 1.7851468889457883e-05,
"loss": 0.5384,
"step": 232
},
{
"epoch": 0.2928112788893907,
"grad_norm": 1.2419851366188785,
"learning_rate": 1.7824181740942958e-05,
"loss": 0.526,
"step": 233
},
{
"epoch": 0.2940679796571563,
"grad_norm": 1.2190078071425596,
"learning_rate": 1.7796743540632226e-05,
"loss": 0.5228,
"step": 234
},
{
"epoch": 0.29532468042492194,
"grad_norm": 1.2620408913566545,
"learning_rate": 1.776915481824107e-05,
"loss": 0.5362,
"step": 235
},
{
"epoch": 0.29658138119268757,
"grad_norm": 1.2019501932459877,
"learning_rate": 1.7741416106390828e-05,
"loss": 0.5403,
"step": 236
},
{
"epoch": 0.2978380819604532,
"grad_norm": 1.2349476994702815,
"learning_rate": 1.7713527940598473e-05,
"loss": 0.5346,
"step": 237
},
{
"epoch": 0.2990947827282188,
"grad_norm": 1.228835848024245,
"learning_rate": 1.7685490859266324e-05,
"loss": 0.5357,
"step": 238
},
{
"epoch": 0.30035148349598445,
"grad_norm": 1.1433816285308742,
"learning_rate": 1.7657305403671618e-05,
"loss": 0.5237,
"step": 239
},
{
"epoch": 0.3016081842637501,
"grad_norm": 1.1830531733373626,
"learning_rate": 1.762897211795607e-05,
"loss": 0.5048,
"step": 240
},
{
"epoch": 0.3028648850315157,
"grad_norm": 1.3012741117414692,
"learning_rate": 1.760049154911537e-05,
"loss": 0.5325,
"step": 241
},
{
"epoch": 0.30412158579928134,
"grad_norm": 1.3019810089595567,
"learning_rate": 1.7571864246988623e-05,
"loss": 0.5311,
"step": 242
},
{
"epoch": 0.30537828656704696,
"grad_norm": 1.2381762150161675,
"learning_rate": 1.7543090764247726e-05,
"loss": 0.5256,
"step": 243
},
{
"epoch": 0.3066349873348126,
"grad_norm": 1.189237177245814,
"learning_rate": 1.751417165638671e-05,
"loss": 0.533,
"step": 244
},
{
"epoch": 0.3078916881025782,
"grad_norm": 1.1392996416893437,
"learning_rate": 1.7485107481711014e-05,
"loss": 0.5116,
"step": 245
},
{
"epoch": 0.30914838887034385,
"grad_norm": 1.2392857084749362,
"learning_rate": 1.7455898801326685e-05,
"loss": 0.5324,
"step": 246
},
{
"epoch": 0.3104050896381095,
"grad_norm": 1.2004003016053595,
"learning_rate": 1.742654617912958e-05,
"loss": 0.5445,
"step": 247
},
{
"epoch": 0.31166179040587505,
"grad_norm": 1.202284432596933,
"learning_rate": 1.7397050181794463e-05,
"loss": 0.5455,
"step": 248
},
{
"epoch": 0.3129184911736407,
"grad_norm": 1.2367358266165964,
"learning_rate": 1.736741137876405e-05,
"loss": 0.5263,
"step": 249
},
{
"epoch": 0.3141751919414063,
"grad_norm": 1.2263686883476443,
"learning_rate": 1.733763034223804e-05,
"loss": 0.5318,
"step": 250
},
{
"epoch": 0.3141751919414063,
"eval_loss": NaN,
"eval_runtime": 393.3082,
"eval_samples_per_second": 21.141,
"eval_steps_per_second": 2.644,
"step": 250
},
{
"epoch": 0.31543189270917193,
"grad_norm": 1.839889882151207,
"learning_rate": 1.730770764716206e-05,
"loss": 0.4856,
"step": 251
},
{
"epoch": 0.31668859347693756,
"grad_norm": 1.2987969186805615,
"learning_rate": 1.7277643871216558e-05,
"loss": 0.5393,
"step": 252
},
{
"epoch": 0.3179452942447032,
"grad_norm": 1.4429089329852214,
"learning_rate": 1.724743959480565e-05,
"loss": 0.5473,
"step": 253
},
{
"epoch": 0.3192019950124688,
"grad_norm": 2.3182483058377357,
"learning_rate": 1.721709540104594e-05,
"loss": 0.5719,
"step": 254
},
{
"epoch": 0.32045869578023445,
"grad_norm": 1.2943046560287874,
"learning_rate": 1.7186611875755227e-05,
"loss": 0.5421,
"step": 255
},
{
"epoch": 0.3217153965480001,
"grad_norm": 1.3312637937305758,
"learning_rate": 1.715598960744121e-05,
"loss": 0.5756,
"step": 256
},
{
"epoch": 0.3229720973157657,
"grad_norm": 1.203432712589985,
"learning_rate": 1.712522918729014e-05,
"loss": 0.5876,
"step": 257
},
{
"epoch": 0.32422879808353133,
"grad_norm": 1.248392549739193,
"learning_rate": 1.7094331209155394e-05,
"loss": 0.5375,
"step": 258
},
{
"epoch": 0.32548549885129696,
"grad_norm": 1.3256545033784541,
"learning_rate": 1.7063296269545988e-05,
"loss": 0.5067,
"step": 259
},
{
"epoch": 0.3267421996190626,
"grad_norm": 1.1874224988844255,
"learning_rate": 1.7032124967615112e-05,
"loss": 0.5281,
"step": 260
},
{
"epoch": 0.3279989003868282,
"grad_norm": 1.36231925710455,
"learning_rate": 1.7000817905148523e-05,
"loss": 0.546,
"step": 261
},
{
"epoch": 0.32925560115459385,
"grad_norm": 1.2674188969413769,
"learning_rate": 1.696937568655294e-05,
"loss": 0.5139,
"step": 262
},
{
"epoch": 0.3305123019223595,
"grad_norm": 1.2676550984339194,
"learning_rate": 1.6937798918844363e-05,
"loss": 0.5033,
"step": 263
},
{
"epoch": 0.3317690026901251,
"grad_norm": 1.2627941544296886,
"learning_rate": 1.6906088211636387e-05,
"loss": 0.5323,
"step": 264
},
{
"epoch": 0.33302570345789073,
"grad_norm": 1.1828449763849782,
"learning_rate": 1.6874244177128395e-05,
"loss": 0.4978,
"step": 265
},
{
"epoch": 0.3342824042256563,
"grad_norm": 1.2113446935401968,
"learning_rate": 1.6842267430093762e-05,
"loss": 0.5336,
"step": 266
},
{
"epoch": 0.33553910499342193,
"grad_norm": 1.1871927265354432,
"learning_rate": 1.6810158587867973e-05,
"loss": 0.4963,
"step": 267
},
{
"epoch": 0.33679580576118756,
"grad_norm": 1.21086248350049,
"learning_rate": 1.6777918270336718e-05,
"loss": 0.5204,
"step": 268
},
{
"epoch": 0.3380525065289532,
"grad_norm": 1.2862931993679727,
"learning_rate": 1.6745547099923917e-05,
"loss": 0.5306,
"step": 269
},
{
"epoch": 0.3393092072967188,
"grad_norm": 1.2086868536312418,
"learning_rate": 1.6713045701579705e-05,
"loss": 0.4775,
"step": 270
},
{
"epoch": 0.34056590806448445,
"grad_norm": 1.246734645659461,
"learning_rate": 1.6680414702768358e-05,
"loss": 0.5003,
"step": 271
},
{
"epoch": 0.3418226088322501,
"grad_norm": 1.1307518995270345,
"learning_rate": 1.66476547334562e-05,
"loss": 0.5618,
"step": 272
},
{
"epoch": 0.3430793096000157,
"grad_norm": 1.2461483743109123,
"learning_rate": 1.661476642609943e-05,
"loss": 0.5154,
"step": 273
},
{
"epoch": 0.34433601036778133,
"grad_norm": 1.2318638718720292,
"learning_rate": 1.658175041563189e-05,
"loss": 0.5255,
"step": 274
},
{
"epoch": 0.34559271113554696,
"grad_norm": 1.2614769482466344,
"learning_rate": 1.6548607339452853e-05,
"loss": 0.4903,
"step": 275
},
{
"epoch": 0.3468494119033126,
"grad_norm": 1.218403889343406,
"learning_rate": 1.6515337837414677e-05,
"loss": 0.5374,
"step": 276
},
{
"epoch": 0.3481061126710782,
"grad_norm": 1.309250396407954,
"learning_rate": 1.6481942551810476e-05,
"loss": 0.5207,
"step": 277
},
{
"epoch": 0.34936281343884384,
"grad_norm": 1.1711784327085948,
"learning_rate": 1.6448422127361707e-05,
"loss": 0.531,
"step": 278
},
{
"epoch": 0.3506195142066095,
"grad_norm": 1.1878422695605408,
"learning_rate": 1.641477721120573e-05,
"loss": 0.54,
"step": 279
},
{
"epoch": 0.3518762149743751,
"grad_norm": 1.161149033672964,
"learning_rate": 1.638100845288331e-05,
"loss": 0.5078,
"step": 280
},
{
"epoch": 0.35313291574214073,
"grad_norm": 1.1916167026440183,
"learning_rate": 1.6347116504326082e-05,
"loss": 0.5246,
"step": 281
},
{
"epoch": 0.35438961650990636,
"grad_norm": 1.1648314835473719,
"learning_rate": 1.631310201984396e-05,
"loss": 0.5111,
"step": 282
},
{
"epoch": 0.355646317277672,
"grad_norm": 1.2652965637545066,
"learning_rate": 1.627896565611251e-05,
"loss": 0.5376,
"step": 283
},
{
"epoch": 0.3569030180454376,
"grad_norm": 1.0925768418862014,
"learning_rate": 1.6244708072160267e-05,
"loss": 0.5158,
"step": 284
},
{
"epoch": 0.3581597188132032,
"grad_norm": 1.237142044854975,
"learning_rate": 1.6210329929356017e-05,
"loss": 0.4975,
"step": 285
},
{
"epoch": 0.3594164195809688,
"grad_norm": 1.314890455752597,
"learning_rate": 1.6175831891396034e-05,
"loss": 0.5504,
"step": 286
},
{
"epoch": 0.36067312034873444,
"grad_norm": 1.2189337274031111,
"learning_rate": 1.614121462429124e-05,
"loss": 0.55,
"step": 287
},
{
"epoch": 0.3619298211165001,
"grad_norm": 1.1195814719052748,
"learning_rate": 1.6106478796354382e-05,
"loss": 0.5159,
"step": 288
},
{
"epoch": 0.3631865218842657,
"grad_norm": 1.2263611549293467,
"learning_rate": 1.6071625078187113e-05,
"loss": 0.5478,
"step": 289
},
{
"epoch": 0.36444322265203133,
"grad_norm": 1.150066595637109,
"learning_rate": 1.6036654142667043e-05,
"loss": 0.5503,
"step": 290
},
{
"epoch": 0.36569992341979696,
"grad_norm": 1.1156487171205896,
"learning_rate": 1.600156666493475e-05,
"loss": 0.5218,
"step": 291
},
{
"epoch": 0.3669566241875626,
"grad_norm": 1.132115800963164,
"learning_rate": 1.596636332238076e-05,
"loss": 0.4936,
"step": 292
},
{
"epoch": 0.3682133249553282,
"grad_norm": 1.1553749691225366,
"learning_rate": 1.593104479463244e-05,
"loss": 0.5209,
"step": 293
},
{
"epoch": 0.36947002572309384,
"grad_norm": 1.1883358272581606,
"learning_rate": 1.5895611763540914e-05,
"loss": 0.4964,
"step": 294
},
{
"epoch": 0.37072672649085947,
"grad_norm": 1.1558054507595528,
"learning_rate": 1.5860064913167863e-05,
"loss": 0.5063,
"step": 295
},
{
"epoch": 0.3719834272586251,
"grad_norm": 1.2438113556876693,
"learning_rate": 1.5824404929772347e-05,
"loss": 0.507,
"step": 296
},
{
"epoch": 0.3732401280263907,
"grad_norm": 1.1179995486577579,
"learning_rate": 1.5788632501797545e-05,
"loss": 0.4902,
"step": 297
},
{
"epoch": 0.37449682879415636,
"grad_norm": 1.1092026267657866,
"learning_rate": 1.575274831985746e-05,
"loss": 0.512,
"step": 298
},
{
"epoch": 0.375753529561922,
"grad_norm": 1.154632467938108,
"learning_rate": 1.5716753076723594e-05,
"loss": 0.5245,
"step": 299
},
{
"epoch": 0.3770102303296876,
"grad_norm": 1.189993649393655,
"learning_rate": 1.568064746731156e-05,
"loss": 0.5191,
"step": 300
},
{
"epoch": 0.3770102303296876,
"eval_loss": NaN,
"eval_runtime": 390.9329,
"eval_samples_per_second": 21.27,
"eval_steps_per_second": 2.66,
"step": 300
},
{
"epoch": 0.37826693109745324,
"grad_norm": 1.1500530686780315,
"learning_rate": 1.5644432188667695e-05,
"loss": 0.5288,
"step": 301
},
{
"epoch": 0.37952363186521887,
"grad_norm": 1.1878804551065247,
"learning_rate": 1.5608107939955565e-05,
"loss": 0.5231,
"step": 302
},
{
"epoch": 0.38078033263298444,
"grad_norm": 1.1168332905551357,
"learning_rate": 1.5571675422442504e-05,
"loss": 0.5387,
"step": 303
},
{
"epoch": 0.38203703340075007,
"grad_norm": 1.1448743390588387,
"learning_rate": 1.5535135339486044e-05,
"loss": 0.5214,
"step": 304
},
{
"epoch": 0.3832937341685157,
"grad_norm": 1.1699623266496315,
"learning_rate": 1.549848839652035e-05,
"loss": 0.5466,
"step": 305
},
{
"epoch": 0.3845504349362813,
"grad_norm": 1.1274856023440132,
"learning_rate": 1.5461735301042615e-05,
"loss": 0.5534,
"step": 306
},
{
"epoch": 0.38580713570404696,
"grad_norm": 1.1200748824733626,
"learning_rate": 1.542487676259937e-05,
"loss": 0.5213,
"step": 307
},
{
"epoch": 0.3870638364718126,
"grad_norm": 1.1177676101897944,
"learning_rate": 1.5387913492772816e-05,
"loss": 0.5261,
"step": 308
},
{
"epoch": 0.3883205372395782,
"grad_norm": 1.0956645557598703,
"learning_rate": 1.5350846205167065e-05,
"loss": 0.5007,
"step": 309
},
{
"epoch": 0.38957723800734384,
"grad_norm": 1.170185855384506,
"learning_rate": 1.5313675615394373e-05,
"loss": 0.5221,
"step": 310
},
{
"epoch": 0.39083393877510947,
"grad_norm": 1.1889108508498927,
"learning_rate": 1.527640244106133e-05,
"loss": 0.5385,
"step": 311
},
{
"epoch": 0.3920906395428751,
"grad_norm": 1.1415465303312236,
"learning_rate": 1.5239027401754987e-05,
"loss": 0.5112,
"step": 312
},
{
"epoch": 0.3933473403106407,
"grad_norm": 1.2013200373878787,
"learning_rate": 1.5201551219028988e-05,
"loss": 0.4914,
"step": 313
},
{
"epoch": 0.39460404107840635,
"grad_norm": 1.1988351313938457,
"learning_rate": 1.5163974616389621e-05,
"loss": 0.4947,
"step": 314
},
{
"epoch": 0.395860741846172,
"grad_norm": 1.089460298631534,
"learning_rate": 1.5126298319281859e-05,
"loss": 0.5458,
"step": 315
},
{
"epoch": 0.3971174426139376,
"grad_norm": 1.1216780572022857,
"learning_rate": 1.508852305507535e-05,
"loss": 0.529,
"step": 316
},
{
"epoch": 0.39837414338170324,
"grad_norm": 1.1750196353813835,
"learning_rate": 1.5050649553050383e-05,
"loss": 0.5343,
"step": 317
},
{
"epoch": 0.39963084414946887,
"grad_norm": 1.1613435161044152,
"learning_rate": 1.5012678544383797e-05,
"loss": 0.4809,
"step": 318
},
{
"epoch": 0.4008875449172345,
"grad_norm": 1.0838357359556312,
"learning_rate": 1.4974610762134875e-05,
"loss": 0.5397,
"step": 319
},
{
"epoch": 0.4021442456850001,
"grad_norm": 1.1334411455279323,
"learning_rate": 1.4936446941231186e-05,
"loss": 0.5023,
"step": 320
},
{
"epoch": 0.4034009464527657,
"grad_norm": 1.084067016750503,
"learning_rate": 1.4898187818454401e-05,
"loss": 0.507,
"step": 321
},
{
"epoch": 0.4046576472205313,
"grad_norm": 1.1436033686465177,
"learning_rate": 1.485983413242606e-05,
"loss": 0.4803,
"step": 322
},
{
"epoch": 0.40591434798829695,
"grad_norm": 1.1381556287788788,
"learning_rate": 1.4821386623593332e-05,
"loss": 0.5212,
"step": 323
},
{
"epoch": 0.4071710487560626,
"grad_norm": 1.2083085101416968,
"learning_rate": 1.4782846034214689e-05,
"loss": 0.6227,
"step": 324
},
{
"epoch": 0.4084277495238282,
"grad_norm": 1.1218491511435251,
"learning_rate": 1.4744213108345605e-05,
"loss": 0.5064,
"step": 325
},
{
"epoch": 0.40968445029159384,
"grad_norm": 1.222423494749568,
"learning_rate": 1.4705488591824182e-05,
"loss": 0.5404,
"step": 326
},
{
"epoch": 0.41094115105935947,
"grad_norm": 1.3200013736515412,
"learning_rate": 1.4666673232256738e-05,
"loss": 0.5236,
"step": 327
},
{
"epoch": 0.4121978518271251,
"grad_norm": 1.174130976702849,
"learning_rate": 1.4627767779003402e-05,
"loss": 0.5109,
"step": 328
},
{
"epoch": 0.4134545525948907,
"grad_norm": 1.1121745715846247,
"learning_rate": 1.4588772983163612e-05,
"loss": 0.5187,
"step": 329
},
{
"epoch": 0.41471125336265635,
"grad_norm": 1.1211595150469165,
"learning_rate": 1.4549689597561652e-05,
"loss": 0.5195,
"step": 330
},
{
"epoch": 0.415967954130422,
"grad_norm": 1.0882830091956517,
"learning_rate": 1.4510518376732081e-05,
"loss": 0.5197,
"step": 331
},
{
"epoch": 0.4172246548981876,
"grad_norm": 1.103233638957626,
"learning_rate": 1.4471260076905198e-05,
"loss": 0.5025,
"step": 332
},
{
"epoch": 0.41848135566595324,
"grad_norm": 1.073960984254828,
"learning_rate": 1.4431915455992416e-05,
"loss": 0.5093,
"step": 333
},
{
"epoch": 0.41973805643371886,
"grad_norm": 1.0782782404257423,
"learning_rate": 1.4392485273571652e-05,
"loss": 0.4834,
"step": 334
},
{
"epoch": 0.4209947572014845,
"grad_norm": 1.1812719353147918,
"learning_rate": 1.435297029087265e-05,
"loss": 0.5096,
"step": 335
},
{
"epoch": 0.4222514579692501,
"grad_norm": 1.0945216590303213,
"learning_rate": 1.431337127076229e-05,
"loss": 0.4992,
"step": 336
},
{
"epoch": 0.42350815873701575,
"grad_norm": 1.124754349061893,
"learning_rate": 1.4273688977729852e-05,
"loss": 0.5806,
"step": 337
},
{
"epoch": 0.4247648595047814,
"grad_norm": 1.182469605262903,
"learning_rate": 1.4233924177872269e-05,
"loss": 0.5119,
"step": 338
},
{
"epoch": 0.426021560272547,
"grad_norm": 1.095989659016963,
"learning_rate": 1.4194077638879333e-05,
"loss": 0.5369,
"step": 339
},
{
"epoch": 0.4272782610403126,
"grad_norm": 1.1287938362303405,
"learning_rate": 1.4154150130018867e-05,
"loss": 0.4763,
"step": 340
},
{
"epoch": 0.4285349618080782,
"grad_norm": 1.1391290234474063,
"learning_rate": 1.4114142422121879e-05,
"loss": 0.5052,
"step": 341
},
{
"epoch": 0.42979166257584384,
"grad_norm": 1.0672829625786657,
"learning_rate": 1.4074055287567685e-05,
"loss": 0.539,
"step": 342
},
{
"epoch": 0.43104836334360946,
"grad_norm": 1.1130997359221613,
"learning_rate": 1.4033889500268991e-05,
"loss": 0.5033,
"step": 343
},
{
"epoch": 0.4323050641113751,
"grad_norm": 1.11023882693686,
"learning_rate": 1.3993645835656955e-05,
"loss": 0.4961,
"step": 344
},
{
"epoch": 0.4335617648791407,
"grad_norm": 1.129690617178472,
"learning_rate": 1.3953325070666215e-05,
"loss": 0.5302,
"step": 345
},
{
"epoch": 0.43481846564690635,
"grad_norm": 1.1112927798263614,
"learning_rate": 1.3912927983719888e-05,
"loss": 0.5032,
"step": 346
},
{
"epoch": 0.436075166414672,
"grad_norm": 1.1287505890600753,
"learning_rate": 1.3872455354714552e-05,
"loss": 0.5287,
"step": 347
},
{
"epoch": 0.4373318671824376,
"grad_norm": 1.1474684694732853,
"learning_rate": 1.3831907965005173e-05,
"loss": 0.503,
"step": 348
},
{
"epoch": 0.43858856795020323,
"grad_norm": 1.1271461572724835,
"learning_rate": 1.3791286597390035e-05,
"loss": 0.52,
"step": 349
},
{
"epoch": 0.43984526871796886,
"grad_norm": 1.0112579946119413,
"learning_rate": 1.375059203609562e-05,
"loss": 0.512,
"step": 350
},
{
"epoch": 0.43984526871796886,
"eval_loss": NaN,
"eval_runtime": 386.8898,
"eval_samples_per_second": 21.492,
"eval_steps_per_second": 2.688,
"step": 350
},
{
"epoch": 0.4411019694857345,
"grad_norm": 1.1391998689599176,
"learning_rate": 1.370982506676147e-05,
"loss": 0.5153,
"step": 351
},
{
"epoch": 0.4423586702535001,
"grad_norm": 1.1522319157530978,
"learning_rate": 1.3668986476425024e-05,
"loss": 0.5223,
"step": 352
},
{
"epoch": 0.44361537102126575,
"grad_norm": 1.056841423994722,
"learning_rate": 1.362807705350641e-05,
"loss": 0.5459,
"step": 353
},
{
"epoch": 0.4448720717890314,
"grad_norm": 1.1383935709291246,
"learning_rate": 1.3587097587793243e-05,
"loss": 0.5449,
"step": 354
},
{
"epoch": 0.446128772556797,
"grad_norm": 1.106248398688068,
"learning_rate": 1.3546048870425356e-05,
"loss": 0.4944,
"step": 355
},
{
"epoch": 0.44738547332456263,
"grad_norm": 1.067289329127628,
"learning_rate": 1.3504931693879553e-05,
"loss": 0.5538,
"step": 356
},
{
"epoch": 0.44864217409232826,
"grad_norm": 1.0795347423113244,
"learning_rate": 1.3463746851954275e-05,
"loss": 0.5306,
"step": 357
},
{
"epoch": 0.44989887486009383,
"grad_norm": 1.1163481113252645,
"learning_rate": 1.342249513975432e-05,
"loss": 0.5146,
"step": 358
},
{
"epoch": 0.45115557562785946,
"grad_norm": 1.1760911501443567,
"learning_rate": 1.3381177353675441e-05,
"loss": 0.5239,
"step": 359
},
{
"epoch": 0.4524122763956251,
"grad_norm": 1.1510645603762788,
"learning_rate": 1.3339794291389015e-05,
"loss": 0.4986,
"step": 360
},
{
"epoch": 0.4536689771633907,
"grad_norm": 1.0505550098423375,
"learning_rate": 1.3298346751826624e-05,
"loss": 0.4932,
"step": 361
},
{
"epoch": 0.45492567793115635,
"grad_norm": 1.110458489656852,
"learning_rate": 1.3256835535164622e-05,
"loss": 0.499,
"step": 362
},
{
"epoch": 0.456182378698922,
"grad_norm": 1.0429297164870697,
"learning_rate": 1.3215261442808718e-05,
"loss": 0.5519,
"step": 363
},
{
"epoch": 0.4574390794666876,
"grad_norm": 1.1025779859161235,
"learning_rate": 1.3173625277378464e-05,
"loss": 0.498,
"step": 364
},
{
"epoch": 0.45869578023445323,
"grad_norm": 1.0866619959147732,
"learning_rate": 1.3131927842691793e-05,
"loss": 0.4939,
"step": 365
},
{
"epoch": 0.45995248100221886,
"grad_norm": 1.0657930951927579,
"learning_rate": 1.3090169943749475e-05,
"loss": 0.4984,
"step": 366
},
{
"epoch": 0.4612091817699845,
"grad_norm": 1.0589212846976253,
"learning_rate": 1.3048352386719605e-05,
"loss": 0.4943,
"step": 367
},
{
"epoch": 0.4624658825377501,
"grad_norm": 1.0942767373527489,
"learning_rate": 1.3006475978922013e-05,
"loss": 0.5145,
"step": 368
},
{
"epoch": 0.46372258330551575,
"grad_norm": 1.123925914892006,
"learning_rate": 1.2964541528812689e-05,
"loss": 0.5212,
"step": 369
},
{
"epoch": 0.4649792840732814,
"grad_norm": 1.1029621728132757,
"learning_rate": 1.2922549845968174e-05,
"loss": 0.5008,
"step": 370
},
{
"epoch": 0.466235984841047,
"grad_norm": 1.1066345430120166,
"learning_rate": 1.2880501741069931e-05,
"loss": 0.5099,
"step": 371
},
{
"epoch": 0.46749268560881263,
"grad_norm": 1.1178828922894442,
"learning_rate": 1.2838398025888695e-05,
"loss": 0.5174,
"step": 372
},
{
"epoch": 0.46874938637657826,
"grad_norm": 1.0618894730289437,
"learning_rate": 1.2796239513268796e-05,
"loss": 0.5266,
"step": 373
},
{
"epoch": 0.4700060871443439,
"grad_norm": 1.1442577048190856,
"learning_rate": 1.275402701711248e-05,
"loss": 0.5406,
"step": 374
},
{
"epoch": 0.4712627879121095,
"grad_norm": 1.5555245911074627,
"learning_rate": 1.2711761352364172e-05,
"loss": 0.526,
"step": 375
},
{
"epoch": 0.4725194886798751,
"grad_norm": 3.2281540462224814,
"learning_rate": 1.2669443334994768e-05,
"loss": 0.5058,
"step": 376
},
{
"epoch": 0.4737761894476407,
"grad_norm": 1.230186696498868,
"learning_rate": 1.262707378198587e-05,
"loss": 0.5189,
"step": 377
},
{
"epoch": 0.47503289021540634,
"grad_norm": 1.1067717727956337,
"learning_rate": 1.2584653511314012e-05,
"loss": 0.5499,
"step": 378
},
{
"epoch": 0.476289590983172,
"grad_norm": 1.2096032452146201,
"learning_rate": 1.2542183341934873e-05,
"loss": 0.4961,
"step": 379
},
{
"epoch": 0.4775462917509376,
"grad_norm": 1.1760777197898722,
"learning_rate": 1.2499664093767458e-05,
"loss": 0.5352,
"step": 380
},
{
"epoch": 0.47880299251870323,
"grad_norm": 1.1829467616609262,
"learning_rate": 1.245709658767829e-05,
"loss": 0.5039,
"step": 381
},
{
"epoch": 0.48005969328646886,
"grad_norm": 1.093348347620876,
"learning_rate": 1.241448164546553e-05,
"loss": 0.5063,
"step": 382
},
{
"epoch": 0.4813163940542345,
"grad_norm": 1.0813316534886166,
"learning_rate": 1.2371820089843145e-05,
"loss": 0.504,
"step": 383
},
{
"epoch": 0.4825730948220001,
"grad_norm": 1.0559642834186367,
"learning_rate": 1.2329112744425e-05,
"loss": 0.4968,
"step": 384
},
{
"epoch": 0.48382979558976574,
"grad_norm": 1.1119456528103215,
"learning_rate": 1.2286360433708976e-05,
"loss": 0.5623,
"step": 385
},
{
"epoch": 0.48508649635753137,
"grad_norm": 1.143973204715742,
"learning_rate": 1.2243563983061029e-05,
"loss": 0.4909,
"step": 386
},
{
"epoch": 0.486343197125297,
"grad_norm": 1.0989207521976259,
"learning_rate": 1.2200724218699284e-05,
"loss": 0.5342,
"step": 387
},
{
"epoch": 0.48759989789306263,
"grad_norm": 1.097054041408541,
"learning_rate": 1.2157841967678064e-05,
"loss": 0.4762,
"step": 388
},
{
"epoch": 0.48885659866082826,
"grad_norm": 1.0459339979949123,
"learning_rate": 1.2114918057871928e-05,
"loss": 0.4651,
"step": 389
},
{
"epoch": 0.4901132994285939,
"grad_norm": 1.0847962784136453,
"learning_rate": 1.2071953317959692e-05,
"loss": 0.5305,
"step": 390
},
{
"epoch": 0.4913700001963595,
"grad_norm": 1.1246107275670814,
"learning_rate": 1.202894857740843e-05,
"loss": 0.524,
"step": 391
},
{
"epoch": 0.49262670096412514,
"grad_norm": 1.0117888555208185,
"learning_rate": 1.1985904666457455e-05,
"loss": 0.4387,
"step": 392
},
{
"epoch": 0.49388340173189077,
"grad_norm": 1.072130705469674,
"learning_rate": 1.19428224161023e-05,
"loss": 0.4996,
"step": 393
},
{
"epoch": 0.4951401024996564,
"grad_norm": 1.0949029023305772,
"learning_rate": 1.1899702658078663e-05,
"loss": 0.5001,
"step": 394
},
{
"epoch": 0.49639680326742197,
"grad_norm": 1.0307577287604857,
"learning_rate": 1.1856546224846354e-05,
"loss": 0.5561,
"step": 395
},
{
"epoch": 0.4976535040351876,
"grad_norm": 1.0754967872326906,
"learning_rate": 1.181335394957324e-05,
"loss": 0.4659,
"step": 396
},
{
"epoch": 0.4989102048029532,
"grad_norm": 1.078728824641922,
"learning_rate": 1.1770126666119133e-05,
"loss": 0.5014,
"step": 397
},
{
"epoch": 0.5001669055707189,
"grad_norm": 1.0499497992004496,
"learning_rate": 1.1726865209019709e-05,
"loss": 0.4716,
"step": 398
},
{
"epoch": 0.5014236063384845,
"grad_norm": 1.0777776407348791,
"learning_rate": 1.1683570413470384e-05,
"loss": 0.4928,
"step": 399
},
{
"epoch": 0.5026803071062501,
"grad_norm": 1.0574879518909284,
"learning_rate": 1.1640243115310219e-05,
"loss": 0.4814,
"step": 400
},
{
"epoch": 0.5026803071062501,
"eval_loss": NaN,
"eval_runtime": 389.4493,
"eval_samples_per_second": 21.351,
"eval_steps_per_second": 2.67,
"step": 400
},
{
"epoch": 0.5039370078740157,
"grad_norm": 1.0747572780095613,
"learning_rate": 1.1596884151005743e-05,
"loss": 0.5086,
"step": 401
},
{
"epoch": 0.5051937086417814,
"grad_norm": 1.104818199875437,
"learning_rate": 1.155349435763483e-05,
"loss": 0.5385,
"step": 402
},
{
"epoch": 0.506450409409547,
"grad_norm": 1.0738096255710292,
"learning_rate": 1.1510074572870533e-05,
"loss": 0.4624,
"step": 403
},
{
"epoch": 0.5077071101773126,
"grad_norm": 1.0669378834870606,
"learning_rate": 1.1466625634964911e-05,
"loss": 0.5015,
"step": 404
},
{
"epoch": 0.5089638109450783,
"grad_norm": 1.0058479382503636,
"learning_rate": 1.1423148382732854e-05,
"loss": 0.4808,
"step": 405
},
{
"epoch": 0.5102205117128439,
"grad_norm": 1.0558622246582776,
"learning_rate": 1.1379643655535869e-05,
"loss": 0.4676,
"step": 406
},
{
"epoch": 0.5114772124806095,
"grad_norm": 1.0593950117803592,
"learning_rate": 1.1336112293265896e-05,
"loss": 0.5675,
"step": 407
},
{
"epoch": 0.5127339132483751,
"grad_norm": 1.0819776501101257,
"learning_rate": 1.1292555136329082e-05,
"loss": 0.523,
"step": 408
},
{
"epoch": 0.5139906140161408,
"grad_norm": 1.1503702106834164,
"learning_rate": 1.1248973025629567e-05,
"loss": 0.573,
"step": 409
},
{
"epoch": 0.5152473147839064,
"grad_norm": 1.1896682820272197,
"learning_rate": 1.1205366802553231e-05,
"loss": 0.5443,
"step": 410
},
{
"epoch": 0.516504015551672,
"grad_norm": 1.051710377423096,
"learning_rate": 1.1161737308951473e-05,
"loss": 0.5213,
"step": 411
},
{
"epoch": 0.5177607163194377,
"grad_norm": 1.0876393327574223,
"learning_rate": 1.111808538712494e-05,
"loss": 0.4783,
"step": 412
},
{
"epoch": 0.5190174170872033,
"grad_norm": 1.1011922768783085,
"learning_rate": 1.1074411879807271e-05,
"loss": 0.5381,
"step": 413
},
{
"epoch": 0.5202741178549689,
"grad_norm": 1.0360462896838518,
"learning_rate": 1.1030717630148839e-05,
"loss": 0.4827,
"step": 414
},
{
"epoch": 0.5215308186227345,
"grad_norm": 1.0634457362167398,
"learning_rate": 1.0987003481700456e-05,
"loss": 0.468,
"step": 415
},
{
"epoch": 0.5227875193905002,
"grad_norm": 1.063398934605818,
"learning_rate": 1.0943270278397097e-05,
"loss": 0.5022,
"step": 416
},
{
"epoch": 0.5240442201582658,
"grad_norm": 1.1236585460615724,
"learning_rate": 1.0899518864541607e-05,
"loss": 0.4927,
"step": 417
},
{
"epoch": 0.5253009209260314,
"grad_norm": 1.1131356127327812,
"learning_rate": 1.08557500847884e-05,
"loss": 0.482,
"step": 418
},
{
"epoch": 0.526557621693797,
"grad_norm": 1.074020533914174,
"learning_rate": 1.0811964784127145e-05,
"loss": 0.4989,
"step": 419
},
{
"epoch": 0.5278143224615627,
"grad_norm": 1.10409363319172,
"learning_rate": 1.076816380786647e-05,
"loss": 0.5241,
"step": 420
},
{
"epoch": 0.5290710232293283,
"grad_norm": 1.122732579577094,
"learning_rate": 1.0724348001617626e-05,
"loss": 0.5156,
"step": 421
},
{
"epoch": 0.5303277239970938,
"grad_norm": 1.0617618240477706,
"learning_rate": 1.0680518211278169e-05,
"loss": 0.48,
"step": 422
},
{
"epoch": 0.5315844247648595,
"grad_norm": 1.0479305127381242,
"learning_rate": 1.063667528301563e-05,
"loss": 0.4979,
"step": 423
},
{
"epoch": 0.5328411255326251,
"grad_norm": 0.9841907268191958,
"learning_rate": 1.0592820063251177e-05,
"loss": 0.4794,
"step": 424
},
{
"epoch": 0.5340978263003907,
"grad_norm": 1.0319821042091273,
"learning_rate": 1.0548953398643276e-05,
"loss": 0.5059,
"step": 425
},
{
"epoch": 0.5353545270681563,
"grad_norm": 1.0138228640258524,
"learning_rate": 1.0505076136071342e-05,
"loss": 0.4951,
"step": 426
},
{
"epoch": 0.536611227835922,
"grad_norm": 1.1594538762080069,
"learning_rate": 1.0461189122619394e-05,
"loss": 0.4968,
"step": 427
},
{
"epoch": 0.5378679286036876,
"grad_norm": 1.007644935103111,
"learning_rate": 1.0417293205559694e-05,
"loss": 0.5192,
"step": 428
},
{
"epoch": 0.5391246293714532,
"grad_norm": 1.0681339563091352,
"learning_rate": 1.0373389232336404e-05,
"loss": 0.516,
"step": 429
},
{
"epoch": 0.5403813301392189,
"grad_norm": 1.0071232530991345,
"learning_rate": 1.0329478050549208e-05,
"loss": 0.4812,
"step": 430
},
{
"epoch": 0.5416380309069845,
"grad_norm": 1.0658183171395332,
"learning_rate": 1.0285560507936962e-05,
"loss": 0.4668,
"step": 431
},
{
"epoch": 0.5428947316747501,
"grad_norm": 0.97242755575935,
"learning_rate": 1.0241637452361323e-05,
"loss": 0.5024,
"step": 432
},
{
"epoch": 0.5441514324425157,
"grad_norm": 1.1113207771581286,
"learning_rate": 1.0197709731790375e-05,
"loss": 0.4847,
"step": 433
},
{
"epoch": 0.5454081332102814,
"grad_norm": 1.0305552020476478,
"learning_rate": 1.0153778194282269e-05,
"loss": 0.5019,
"step": 434
},
{
"epoch": 0.546664833978047,
"grad_norm": 1.0370594881922448,
"learning_rate": 1.0109843687968837e-05,
"loss": 0.4958,
"step": 435
},
{
"epoch": 0.5479215347458126,
"grad_norm": 1.01930158227936,
"learning_rate": 1.0065907061039234e-05,
"loss": 0.5096,
"step": 436
},
{
"epoch": 0.5491782355135783,
"grad_norm": 0.9871603297074679,
"learning_rate": 1.0021969161723555e-05,
"loss": 0.4685,
"step": 437
},
{
"epoch": 0.5504349362813439,
"grad_norm": 1.0920379317965205,
"learning_rate": 9.97803083827645e-06,
"loss": 0.4644,
"step": 438
},
{
"epoch": 0.5516916370491095,
"grad_norm": 1.054600973465295,
"learning_rate": 9.934092938960769e-06,
"loss": 0.4893,
"step": 439
},
{
"epoch": 0.5529483378168751,
"grad_norm": 1.0201582460505152,
"learning_rate": 9.890156312031165e-06,
"loss": 0.4955,
"step": 440
},
{
"epoch": 0.5542050385846408,
"grad_norm": 0.9716441699106971,
"learning_rate": 9.846221805717734e-06,
"loss": 0.495,
"step": 441
},
{
"epoch": 0.5554617393524064,
"grad_norm": 1.082312481961198,
"learning_rate": 9.802290268209627e-06,
"loss": 0.5257,
"step": 442
},
{
"epoch": 0.556718440120172,
"grad_norm": 1.0097621423303396,
"learning_rate": 9.75836254763868e-06,
"loss": 0.5063,
"step": 443
},
{
"epoch": 0.5579751408879376,
"grad_norm": 1.0275627453990908,
"learning_rate": 9.71443949206304e-06,
"loss": 0.5461,
"step": 444
},
{
"epoch": 0.5592318416557033,
"grad_norm": 0.990883428276697,
"learning_rate": 9.670521949450793e-06,
"loss": 0.4691,
"step": 445
},
{
"epoch": 0.5604885424234689,
"grad_norm": 1.3593079122891867,
"learning_rate": 9.6266107676636e-06,
"loss": 0.4809,
"step": 446
},
{
"epoch": 0.5617452431912345,
"grad_norm": 1.0402512025062964,
"learning_rate": 9.58270679444031e-06,
"loss": 0.5249,
"step": 447
},
{
"epoch": 0.5630019439590002,
"grad_norm": 1.0374899415597172,
"learning_rate": 9.538810877380611e-06,
"loss": 0.4911,
"step": 448
},
{
"epoch": 0.5642586447267658,
"grad_norm": 1.0181963135605545,
"learning_rate": 9.49492386392866e-06,
"loss": 0.5074,
"step": 449
},
{
"epoch": 0.5655153454945314,
"grad_norm": 1.2455466094666068,
"learning_rate": 9.451046601356725e-06,
"loss": 0.5031,
"step": 450
},
{
"epoch": 0.5655153454945314,
"eval_loss": NaN,
"eval_runtime": 390.362,
"eval_samples_per_second": 21.301,
"eval_steps_per_second": 2.664,
"step": 450
},
{
"epoch": 0.566772046262297,
"grad_norm": 1.087941249032518,
"learning_rate": 9.407179936748827e-06,
"loss": 0.4838,
"step": 451
},
{
"epoch": 0.5680287470300627,
"grad_norm": 1.0487351474005773,
"learning_rate": 9.363324716984375e-06,
"loss": 0.6147,
"step": 452
},
{
"epoch": 0.5692854477978283,
"grad_norm": 1.0006830147884598,
"learning_rate": 9.319481788721833e-06,
"loss": 0.4864,
"step": 453
},
{
"epoch": 0.5705421485655939,
"grad_norm": 1.0804045530543167,
"learning_rate": 9.275651998382377e-06,
"loss": 0.465,
"step": 454
},
{
"epoch": 0.5717988493333596,
"grad_norm": 1.0428355842758288,
"learning_rate": 9.231836192133532e-06,
"loss": 0.4821,
"step": 455
},
{
"epoch": 0.5730555501011252,
"grad_norm": 1.0610803769161647,
"learning_rate": 9.188035215872858e-06,
"loss": 0.4867,
"step": 456
},
{
"epoch": 0.5743122508688908,
"grad_norm": 1.0520660600408804,
"learning_rate": 9.144249915211605e-06,
"loss": 0.5214,
"step": 457
},
{
"epoch": 0.5755689516366564,
"grad_norm": 1.0373635995203723,
"learning_rate": 9.100481135458393e-06,
"loss": 0.4943,
"step": 458
},
{
"epoch": 0.576825652404422,
"grad_norm": 1.3517258385540856,
"learning_rate": 9.056729721602904e-06,
"loss": 0.6224,
"step": 459
},
{
"epoch": 0.5780823531721876,
"grad_norm": 1.0301066331386723,
"learning_rate": 9.012996518299547e-06,
"loss": 0.5092,
"step": 460
},
{
"epoch": 0.5793390539399532,
"grad_norm": 1.0099912645887554,
"learning_rate": 8.969282369851163e-06,
"loss": 0.4901,
"step": 461
},
{
"epoch": 0.5805957547077188,
"grad_norm": 1.0482057682006019,
"learning_rate": 8.92558812019273e-06,
"loss": 0.4793,
"step": 462
},
{
"epoch": 0.5818524554754845,
"grad_norm": 1.0475419610115655,
"learning_rate": 8.881914612875062e-06,
"loss": 0.48,
"step": 463
},
{
"epoch": 0.5831091562432501,
"grad_norm": 1.0523184837700892,
"learning_rate": 8.838262691048529e-06,
"loss": 0.4852,
"step": 464
},
{
"epoch": 0.5843658570110157,
"grad_norm": 1.3520200098351463,
"learning_rate": 8.79463319744677e-06,
"loss": 0.4807,
"step": 465
},
{
"epoch": 0.5856225577787814,
"grad_norm": 1.070883661168547,
"learning_rate": 8.751026974370438e-06,
"loss": 0.5023,
"step": 466
},
{
"epoch": 0.586879258546547,
"grad_norm": 1.064553688268609,
"learning_rate": 8.70744486367092e-06,
"loss": 0.4924,
"step": 467
},
{
"epoch": 0.5881359593143126,
"grad_norm": 1.0759932963506342,
"learning_rate": 8.66388770673411e-06,
"loss": 0.4971,
"step": 468
},
{
"epoch": 0.5893926600820782,
"grad_norm": 0.9719685034220976,
"learning_rate": 8.620356344464135e-06,
"loss": 0.4932,
"step": 469
},
{
"epoch": 0.5906493608498439,
"grad_norm": 1.0215732980684333,
"learning_rate": 8.576851617267151e-06,
"loss": 0.4901,
"step": 470
},
{
"epoch": 0.5919060616176095,
"grad_norm": 1.0057794928079793,
"learning_rate": 8.533374365035089e-06,
"loss": 0.4755,
"step": 471
},
{
"epoch": 0.5931627623853751,
"grad_norm": 1.0731603783900507,
"learning_rate": 8.489925427129469e-06,
"loss": 0.5194,
"step": 472
},
{
"epoch": 0.5944194631531408,
"grad_norm": 1.0413553281269248,
"learning_rate": 8.446505642365174e-06,
"loss": 0.4651,
"step": 473
},
{
"epoch": 0.5956761639209064,
"grad_norm": 1.0439372619946545,
"learning_rate": 8.40311584899426e-06,
"loss": 0.5009,
"step": 474
},
{
"epoch": 0.596932864688672,
"grad_norm": 1.079306190387013,
"learning_rate": 8.359756884689785e-06,
"loss": 0.4643,
"step": 475
},
{
"epoch": 0.5981895654564376,
"grad_norm": 0.9998675795140389,
"learning_rate": 8.316429586529616e-06,
"loss": 0.4811,
"step": 476
},
{
"epoch": 0.5994462662242033,
"grad_norm": 0.9612157155286085,
"learning_rate": 8.273134790980295e-06,
"loss": 0.4732,
"step": 477
},
{
"epoch": 0.6007029669919689,
"grad_norm": 1.0216579545060833,
"learning_rate": 8.22987333388087e-06,
"loss": 0.4878,
"step": 478
},
{
"epoch": 0.6019596677597345,
"grad_norm": 1.035592271636439,
"learning_rate": 8.186646050426763e-06,
"loss": 0.4763,
"step": 479
},
{
"epoch": 0.6032163685275002,
"grad_norm": 0.9266009986946163,
"learning_rate": 8.143453775153646e-06,
"loss": 0.4962,
"step": 480
},
{
"epoch": 0.6044730692952658,
"grad_norm": 1.0063210029903829,
"learning_rate": 8.100297341921342e-06,
"loss": 0.5457,
"step": 481
},
{
"epoch": 0.6057297700630314,
"grad_norm": 1.0212142908210302,
"learning_rate": 8.057177583897704e-06,
"loss": 0.4623,
"step": 482
},
{
"epoch": 0.606986470830797,
"grad_norm": 1.0319718233335433,
"learning_rate": 8.014095333542548e-06,
"loss": 0.5095,
"step": 483
},
{
"epoch": 0.6082431715985627,
"grad_norm": 0.9974523217893733,
"learning_rate": 7.971051422591571e-06,
"loss": 0.5192,
"step": 484
},
{
"epoch": 0.6094998723663283,
"grad_norm": 0.9326473530788999,
"learning_rate": 7.928046682040311e-06,
"loss": 0.4574,
"step": 485
},
{
"epoch": 0.6107565731340939,
"grad_norm": 0.9393968466161859,
"learning_rate": 7.885081942128074e-06,
"loss": 0.5029,
"step": 486
},
{
"epoch": 0.6120132739018596,
"grad_norm": 0.9680572001368599,
"learning_rate": 7.84215803232194e-06,
"loss": 0.514,
"step": 487
},
{
"epoch": 0.6132699746696252,
"grad_norm": 1.0704719257655635,
"learning_rate": 7.79927578130072e-06,
"loss": 0.4928,
"step": 488
},
{
"epoch": 0.6145266754373908,
"grad_norm": 1.0200776137947951,
"learning_rate": 7.756436016938973e-06,
"loss": 0.4869,
"step": 489
},
{
"epoch": 0.6157833762051564,
"grad_norm": 0.9387956884468813,
"learning_rate": 7.713639566291028e-06,
"loss": 0.4799,
"step": 490
},
{
"epoch": 0.6170400769729221,
"grad_norm": 0.9312805823610483,
"learning_rate": 7.670887255575003e-06,
"loss": 0.4931,
"step": 491
},
{
"epoch": 0.6182967777406877,
"grad_norm": 0.9295388110852509,
"learning_rate": 7.628179910156859e-06,
"loss": 0.5234,
"step": 492
},
{
"epoch": 0.6195534785084533,
"grad_norm": 0.9969654532505003,
"learning_rate": 7.585518354534473e-06,
"loss": 0.5349,
"step": 493
},
{
"epoch": 0.620810179276219,
"grad_norm": 0.969918008367211,
"learning_rate": 7.542903412321714e-06,
"loss": 0.4843,
"step": 494
},
{
"epoch": 0.6220668800439845,
"grad_norm": 0.9449769047029168,
"learning_rate": 7.500335906232544e-06,
"loss": 0.4639,
"step": 495
},
{
"epoch": 0.6233235808117501,
"grad_norm": 1.033777318960197,
"learning_rate": 7.4578166580651335e-06,
"loss": 0.504,
"step": 496
},
{
"epoch": 0.6245802815795157,
"grad_norm": 0.9914352689303235,
"learning_rate": 7.4153464886859925e-06,
"loss": 0.4981,
"step": 497
},
{
"epoch": 0.6258369823472814,
"grad_norm": 1.056953990111743,
"learning_rate": 7.372926218014131e-06,
"loss": 0.516,
"step": 498
},
{
"epoch": 0.627093683115047,
"grad_norm": 0.9687789789603087,
"learning_rate": 7.330556665005235e-06,
"loss": 0.4599,
"step": 499
},
{
"epoch": 0.6283503838828126,
"grad_norm": 1.0066721178900888,
"learning_rate": 7.2882386476358304e-06,
"loss": 0.4614,
"step": 500
},
{
"epoch": 0.6283503838828126,
"eval_loss": NaN,
"eval_runtime": 384.2451,
"eval_samples_per_second": 21.64,
"eval_steps_per_second": 2.707,
"step": 500
},
{
"epoch": 0.6296070846505782,
"grad_norm": 1.0360193648866638,
"learning_rate": 7.2459729828875256e-06,
"loss": 0.4916,
"step": 501
},
{
"epoch": 0.6308637854183439,
"grad_norm": 0.9139912569370758,
"learning_rate": 7.203760486731204e-06,
"loss": 0.4688,
"step": 502
},
{
"epoch": 0.6321204861861095,
"grad_norm": 1.0077884286770724,
"learning_rate": 7.161601974111308e-06,
"loss": 0.4936,
"step": 503
},
{
"epoch": 0.6333771869538751,
"grad_norm": 0.9993486667429677,
"learning_rate": 7.119498258930073e-06,
"loss": 0.5179,
"step": 504
},
{
"epoch": 0.6346338877216408,
"grad_norm": 0.9964058480599178,
"learning_rate": 7.0774501540318305e-06,
"loss": 0.4745,
"step": 505
},
{
"epoch": 0.6358905884894064,
"grad_norm": 0.9646168801150186,
"learning_rate": 7.035458471187312e-06,
"loss": 0.4865,
"step": 506
},
{
"epoch": 0.637147289257172,
"grad_norm": 0.9977812292191052,
"learning_rate": 6.993524021077989e-06,
"loss": 0.5161,
"step": 507
},
{
"epoch": 0.6384039900249376,
"grad_norm": 0.9788000489126177,
"learning_rate": 6.951647613280397e-06,
"loss": 0.5361,
"step": 508
},
{
"epoch": 0.6396606907927033,
"grad_norm": 0.9186321053151645,
"learning_rate": 6.909830056250527e-06,
"loss": 0.4871,
"step": 509
},
{
"epoch": 0.6409173915604689,
"grad_norm": 1.0035904562057332,
"learning_rate": 6.868072157308213e-06,
"loss": 0.5202,
"step": 510
},
{
"epoch": 0.6421740923282345,
"grad_norm": 0.9810776358670968,
"learning_rate": 6.826374722621536e-06,
"loss": 0.4599,
"step": 511
},
{
"epoch": 0.6434307930960002,
"grad_norm": 0.9283679926638487,
"learning_rate": 6.784738557191284e-06,
"loss": 0.4628,
"step": 512
},
{
"epoch": 0.6446874938637658,
"grad_norm": 0.9994992253986084,
"learning_rate": 6.7431644648353785e-06,
"loss": 0.4674,
"step": 513
},
{
"epoch": 0.6459441946315314,
"grad_norm": 0.9106352041140137,
"learning_rate": 6.701653248173382e-06,
"loss": 0.4625,
"step": 514
},
{
"epoch": 0.647200895399297,
"grad_norm": 0.9500901417881573,
"learning_rate": 6.660205708610987e-06,
"loss": 0.4797,
"step": 515
},
{
"epoch": 0.6484575961670627,
"grad_norm": 0.9472535122755693,
"learning_rate": 6.618822646324563e-06,
"loss": 0.4669,
"step": 516
},
{
"epoch": 0.6497142969348283,
"grad_norm": 0.9589653557097556,
"learning_rate": 6.577504860245684e-06,
"loss": 0.5408,
"step": 517
},
{
"epoch": 0.6509709977025939,
"grad_norm": 0.96979303506713,
"learning_rate": 6.536253148045726e-06,
"loss": 0.4897,
"step": 518
},
{
"epoch": 0.6522276984703596,
"grad_norm": 0.9496625953614718,
"learning_rate": 6.495068306120452e-06,
"loss": 0.4896,
"step": 519
},
{
"epoch": 0.6534843992381252,
"grad_norm": 0.9999476061446844,
"learning_rate": 6.453951129574644e-06,
"loss": 0.5023,
"step": 520
},
{
"epoch": 0.6547411000058908,
"grad_norm": 0.9223335871200387,
"learning_rate": 6.41290241220676e-06,
"loss": 0.5464,
"step": 521
},
{
"epoch": 0.6559978007736564,
"grad_norm": 1.0030864529913075,
"learning_rate": 6.3719229464935915e-06,
"loss": 0.4778,
"step": 522
},
{
"epoch": 0.6572545015414221,
"grad_norm": 0.9309808406108107,
"learning_rate": 6.331013523574978e-06,
"loss": 0.5015,
"step": 523
},
{
"epoch": 0.6585112023091877,
"grad_norm": 1.1823052168866466,
"learning_rate": 6.290174933238531e-06,
"loss": 0.4782,
"step": 524
},
{
"epoch": 0.6597679030769533,
"grad_norm": 0.9878321294333653,
"learning_rate": 6.249407963904381e-06,
"loss": 0.5033,
"step": 525
},
{
"epoch": 0.661024603844719,
"grad_norm": 0.9465968966854122,
"learning_rate": 6.208713402609968e-06,
"loss": 0.4765,
"step": 526
},
{
"epoch": 0.6622813046124846,
"grad_norm": 1.042662157579056,
"learning_rate": 6.168092034994832e-06,
"loss": 0.4847,
"step": 527
},
{
"epoch": 0.6635380053802502,
"grad_norm": 0.9634576640584912,
"learning_rate": 6.127544645285448e-06,
"loss": 0.4634,
"step": 528
},
{
"epoch": 0.6647947061480158,
"grad_norm": 0.9603382195784586,
"learning_rate": 6.087072016280111e-06,
"loss": 0.4902,
"step": 529
},
{
"epoch": 0.6660514069157815,
"grad_norm": 0.9598750622234282,
"learning_rate": 6.046674929333787e-06,
"loss": 0.4806,
"step": 530
},
{
"epoch": 0.6673081076835471,
"grad_norm": 0.9029178527016823,
"learning_rate": 6.006354164343047e-06,
"loss": 0.4577,
"step": 531
},
{
"epoch": 0.6685648084513126,
"grad_norm": 0.9306648630657135,
"learning_rate": 5.96611049973101e-06,
"loss": 0.578,
"step": 532
},
{
"epoch": 0.6698215092190782,
"grad_norm": 0.9745393471796585,
"learning_rate": 5.925944712432317e-06,
"loss": 0.4582,
"step": 533
},
{
"epoch": 0.6710782099868439,
"grad_norm": 0.9854308021827322,
"learning_rate": 5.885857577878122e-06,
"loss": 0.4728,
"step": 534
},
{
"epoch": 0.6723349107546095,
"grad_norm": 0.9505828662992784,
"learning_rate": 5.845849869981137e-06,
"loss": 0.4732,
"step": 535
},
{
"epoch": 0.6735916115223751,
"grad_norm": 0.971008498361879,
"learning_rate": 5.8059223611206716e-06,
"loss": 0.4782,
"step": 536
},
{
"epoch": 0.6748483122901408,
"grad_norm": 1.0199269663151482,
"learning_rate": 5.766075822127735e-06,
"loss": 0.4664,
"step": 537
},
{
"epoch": 0.6761050130579064,
"grad_norm": 0.9414292961876951,
"learning_rate": 5.726311022270152e-06,
"loss": 0.4869,
"step": 538
},
{
"epoch": 0.677361713825672,
"grad_norm": 0.9637472122488508,
"learning_rate": 5.686628729237713e-06,
"loss": 0.4853,
"step": 539
},
{
"epoch": 0.6786184145934376,
"grad_norm": 1.004859916641192,
"learning_rate": 5.647029709127355e-06,
"loss": 0.4755,
"step": 540
},
{
"epoch": 0.6798751153612033,
"grad_norm": 1.0313280249469097,
"learning_rate": 5.6075147264283526e-06,
"loss": 0.4728,
"step": 541
},
{
"epoch": 0.6811318161289689,
"grad_norm": 0.9536505641573378,
"learning_rate": 5.5680845440075885e-06,
"loss": 0.473,
"step": 542
},
{
"epoch": 0.6823885168967345,
"grad_norm": 0.9540350587248954,
"learning_rate": 5.528739923094806e-06,
"loss": 0.4657,
"step": 543
},
{
"epoch": 0.6836452176645001,
"grad_norm": 0.9072196968401911,
"learning_rate": 5.4894816232679195e-06,
"loss": 0.4543,
"step": 544
},
{
"epoch": 0.6849019184322658,
"grad_norm": 0.9814435512849559,
"learning_rate": 5.450310402438353e-06,
"loss": 0.4897,
"step": 545
},
{
"epoch": 0.6861586192000314,
"grad_norm": 0.9601361725013989,
"learning_rate": 5.4112270168363854e-06,
"loss": 0.4749,
"step": 546
},
{
"epoch": 0.687415319967797,
"grad_norm": 0.9663745304412944,
"learning_rate": 5.3722322209966024e-06,
"loss": 0.4848,
"step": 547
},
{
"epoch": 0.6886720207355627,
"grad_norm": 1.0061563886612799,
"learning_rate": 5.333326767743263e-06,
"loss": 0.4673,
"step": 548
},
{
"epoch": 0.6899287215033283,
"grad_norm": 0.9439604050992346,
"learning_rate": 5.294511408175825e-06,
"loss": 0.4853,
"step": 549
},
{
"epoch": 0.6911854222710939,
"grad_norm": 0.962910340515467,
"learning_rate": 5.2557868916543996e-06,
"loss": 0.4791,
"step": 550
},
{
"epoch": 0.6911854222710939,
"eval_loss": NaN,
"eval_runtime": 395.5216,
"eval_samples_per_second": 21.023,
"eval_steps_per_second": 2.629,
"step": 550
},
{
"epoch": 0.6924421230388595,
"grad_norm": 0.9707818718943069,
"learning_rate": 5.217153965785315e-06,
"loss": 0.5109,
"step": 551
},
{
"epoch": 0.6936988238066252,
"grad_norm": 0.9595264837099737,
"learning_rate": 5.178613376406672e-06,
"loss": 0.4675,
"step": 552
},
{
"epoch": 0.6949555245743908,
"grad_norm": 0.9611450529414026,
"learning_rate": 5.14016586757394e-06,
"loss": 0.4814,
"step": 553
},
{
"epoch": 0.6962122253421564,
"grad_norm": 1.0045400911821167,
"learning_rate": 5.1018121815456045e-06,
"loss": 0.5006,
"step": 554
},
{
"epoch": 0.6974689261099221,
"grad_norm": 0.964170058570595,
"learning_rate": 5.063553058768814e-06,
"loss": 0.4615,
"step": 555
},
{
"epoch": 0.6987256268776877,
"grad_norm": 0.974630657567601,
"learning_rate": 5.025389237865128e-06,
"loss": 0.6329,
"step": 556
},
{
"epoch": 0.6999823276454533,
"grad_norm": 0.9940742215685975,
"learning_rate": 4.987321455616206e-06,
"loss": 0.4603,
"step": 557
},
{
"epoch": 0.701239028413219,
"grad_norm": 0.8894890015293643,
"learning_rate": 4.9493504469496235e-06,
"loss": 0.4553,
"step": 558
},
{
"epoch": 0.7024957291809846,
"grad_norm": 0.9992618220443336,
"learning_rate": 4.911476944924651e-06,
"loss": 0.4864,
"step": 559
},
{
"epoch": 0.7037524299487502,
"grad_norm": 0.9141872027503535,
"learning_rate": 4.873701680718146e-06,
"loss": 0.5114,
"step": 560
},
{
"epoch": 0.7050091307165158,
"grad_norm": 0.9326203191415026,
"learning_rate": 4.836025383610382e-06,
"loss": 0.4764,
"step": 561
},
{
"epoch": 0.7062658314842815,
"grad_norm": 0.925599543126751,
"learning_rate": 4.798448780971013e-06,
"loss": 0.4841,
"step": 562
},
{
"epoch": 0.7075225322520471,
"grad_norm": 0.9315828509380828,
"learning_rate": 4.7609725982450176e-06,
"loss": 0.4739,
"step": 563
},
{
"epoch": 0.7087792330198127,
"grad_norm": 0.934793953759224,
"learning_rate": 4.7235975589386715e-06,
"loss": 0.4464,
"step": 564
},
{
"epoch": 0.7100359337875783,
"grad_norm": 0.9341913967846289,
"learning_rate": 4.686324384605629e-06,
"loss": 0.4559,
"step": 565
},
{
"epoch": 0.711292634555344,
"grad_norm": 0.9828397495391764,
"learning_rate": 4.649153794832939e-06,
"loss": 0.5193,
"step": 566
},
{
"epoch": 0.7125493353231096,
"grad_norm": 0.9608848681418946,
"learning_rate": 4.612086507227186e-06,
"loss": 0.4675,
"step": 567
},
{
"epoch": 0.7138060360908752,
"grad_norm": 0.9250968302486778,
"learning_rate": 4.5751232374006304e-06,
"loss": 0.4469,
"step": 568
},
{
"epoch": 0.7150627368586407,
"grad_norm": 0.8931095869951399,
"learning_rate": 4.538264698957387e-06,
"loss": 0.4384,
"step": 569
},
{
"epoch": 0.7163194376264064,
"grad_norm": 0.8957255025032255,
"learning_rate": 4.501511603479653e-06,
"loss": 0.4595,
"step": 570
},
{
"epoch": 0.717576138394172,
"grad_norm": 0.9797783050556119,
"learning_rate": 4.4648646605139605e-06,
"loss": 0.4803,
"step": 571
},
{
"epoch": 0.7188328391619376,
"grad_norm": 1.0477043965745412,
"learning_rate": 4.428324577557501e-06,
"loss": 0.5001,
"step": 572
},
{
"epoch": 0.7200895399297033,
"grad_norm": 0.9402442503704912,
"learning_rate": 4.391892060044435e-06,
"loss": 0.5285,
"step": 573
},
{
"epoch": 0.7213462406974689,
"grad_norm": 0.895533404301798,
"learning_rate": 4.355567811332311e-06,
"loss": 0.4249,
"step": 574
},
{
"epoch": 0.7226029414652345,
"grad_norm": 0.9375812495309651,
"learning_rate": 4.319352532688444e-06,
"loss": 0.4864,
"step": 575
},
{
"epoch": 0.7238596422330001,
"grad_norm": 0.9535540256167436,
"learning_rate": 4.283246923276411e-06,
"loss": 0.451,
"step": 576
},
{
"epoch": 0.7251163430007658,
"grad_norm": 0.9843964347322175,
"learning_rate": 4.247251680142542e-06,
"loss": 0.4934,
"step": 577
},
{
"epoch": 0.7263730437685314,
"grad_norm": 0.902113121251565,
"learning_rate": 4.211367498202456e-06,
"loss": 0.4859,
"step": 578
},
{
"epoch": 0.727629744536297,
"grad_norm": 0.9071019213314944,
"learning_rate": 4.175595070227655e-06,
"loss": 0.463,
"step": 579
},
{
"epoch": 0.7288864453040627,
"grad_norm": 0.9445861043217729,
"learning_rate": 4.13993508683214e-06,
"loss": 0.4639,
"step": 580
},
{
"epoch": 0.7301431460718283,
"grad_norm": 0.9529191091817809,
"learning_rate": 4.1043882364590895e-06,
"loss": 0.458,
"step": 581
},
{
"epoch": 0.7313998468395939,
"grad_norm": 0.943614064462381,
"learning_rate": 4.068955205367559e-06,
"loss": 0.4501,
"step": 582
},
{
"epoch": 0.7326565476073595,
"grad_norm": 0.9501101229235286,
"learning_rate": 4.033636677619242e-06,
"loss": 0.4507,
"step": 583
},
{
"epoch": 0.7339132483751252,
"grad_norm": 0.9696808926218615,
"learning_rate": 3.998433335065251e-06,
"loss": 0.5016,
"step": 584
},
{
"epoch": 0.7351699491428908,
"grad_norm": 0.8737082824243725,
"learning_rate": 3.96334585733296e-06,
"loss": 0.418,
"step": 585
},
{
"epoch": 0.7364266499106564,
"grad_norm": 0.9176906083149247,
"learning_rate": 3.9283749218128885e-06,
"loss": 0.479,
"step": 586
},
{
"epoch": 0.7376833506784221,
"grad_norm": 0.9192555275886508,
"learning_rate": 3.893521203645618e-06,
"loss": 0.49,
"step": 587
},
{
"epoch": 0.7389400514461877,
"grad_norm": 0.9259761814325044,
"learning_rate": 3.858785375708764e-06,
"loss": 0.4658,
"step": 588
},
{
"epoch": 0.7401967522139533,
"grad_norm": 0.8762077783379093,
"learning_rate": 3.824168108603971e-06,
"loss": 0.4825,
"step": 589
},
{
"epoch": 0.7414534529817189,
"grad_norm": 0.9928947391135122,
"learning_rate": 3.7896700706439826e-06,
"loss": 0.5519,
"step": 590
},
{
"epoch": 0.7427101537494846,
"grad_norm": 0.9049374842785678,
"learning_rate": 3.7552919278397335e-06,
"loss": 0.4518,
"step": 591
},
{
"epoch": 0.7439668545172502,
"grad_norm": 0.8936620428913575,
"learning_rate": 3.7210343438874917e-06,
"loss": 0.4341,
"step": 592
},
{
"epoch": 0.7452235552850158,
"grad_norm": 0.9409320634262343,
"learning_rate": 3.6868979801560443e-06,
"loss": 0.4719,
"step": 593
},
{
"epoch": 0.7464802560527815,
"grad_norm": 0.9851785740451936,
"learning_rate": 3.6528834956739224e-06,
"loss": 0.429,
"step": 594
},
{
"epoch": 0.7477369568205471,
"grad_norm": 0.9952385017039359,
"learning_rate": 3.6189915471166927e-06,
"loss": 0.474,
"step": 595
},
{
"epoch": 0.7489936575883127,
"grad_norm": 0.9175118141037356,
"learning_rate": 3.5852227887942713e-06,
"loss": 0.4453,
"step": 596
},
{
"epoch": 0.7502503583560783,
"grad_norm": 0.8948744735918709,
"learning_rate": 3.5515778726382967e-06,
"loss": 0.491,
"step": 597
},
{
"epoch": 0.751507059123844,
"grad_norm": 0.9493043335059733,
"learning_rate": 3.518057448189527e-06,
"loss": 0.4812,
"step": 598
},
{
"epoch": 0.7527637598916096,
"grad_norm": 0.9526119414704524,
"learning_rate": 3.4846621625853248e-06,
"loss": 0.4758,
"step": 599
},
{
"epoch": 0.7540204606593752,
"grad_norm": 0.9375798827152386,
"learning_rate": 3.4513926605471504e-06,
"loss": 0.4703,
"step": 600
},
{
"epoch": 0.7540204606593752,
"eval_loss": NaN,
"eval_runtime": 391.2661,
"eval_samples_per_second": 21.252,
"eval_steps_per_second": 2.658,
"step": 600
},
{
"epoch": 0.7552771614271409,
"grad_norm": 0.9276644067982343,
"learning_rate": 3.4182495843681117e-06,
"loss": 0.461,
"step": 601
},
{
"epoch": 0.7565338621949065,
"grad_norm": 0.9573184146956152,
"learning_rate": 3.385233573900576e-06,
"loss": 0.4812,
"step": 602
},
{
"epoch": 0.7577905629626721,
"grad_norm": 0.9005304918861501,
"learning_rate": 3.3523452665438004e-06,
"loss": 0.4603,
"step": 603
},
{
"epoch": 0.7590472637304377,
"grad_norm": 0.8844651813155415,
"learning_rate": 3.3195852972316435e-06,
"loss": 0.4783,
"step": 604
},
{
"epoch": 0.7603039644982033,
"grad_norm": 0.9053279526971507,
"learning_rate": 3.2869542984202974e-06,
"loss": 0.5253,
"step": 605
},
{
"epoch": 0.7615606652659689,
"grad_norm": 0.995160552063608,
"learning_rate": 3.254452900076083e-06,
"loss": 0.5082,
"step": 606
},
{
"epoch": 0.7628173660337345,
"grad_norm": 0.9879909586608571,
"learning_rate": 3.2220817296632845e-06,
"loss": 0.4499,
"step": 607
},
{
"epoch": 0.7640740668015001,
"grad_norm": 0.9345160999737937,
"learning_rate": 3.1898414121320277e-06,
"loss": 0.4915,
"step": 608
},
{
"epoch": 0.7653307675692658,
"grad_norm": 0.9546540611891579,
"learning_rate": 3.1577325699062424e-06,
"loss": 0.4864,
"step": 609
},
{
"epoch": 0.7665874683370314,
"grad_norm": 0.9216552084428411,
"learning_rate": 3.125755822871607e-06,
"loss": 0.5028,
"step": 610
},
{
"epoch": 0.767844169104797,
"grad_norm": 0.921600882945863,
"learning_rate": 3.093911788363617e-06,
"loss": 0.4773,
"step": 611
},
{
"epoch": 0.7691008698725627,
"grad_norm": 0.9429114879803641,
"learning_rate": 3.062201081155637e-06,
"loss": 0.4828,
"step": 612
},
{
"epoch": 0.7703575706403283,
"grad_norm": 0.9468558457454733,
"learning_rate": 3.0306243134470668e-06,
"loss": 0.5299,
"step": 613
},
{
"epoch": 0.7716142714080939,
"grad_norm": 0.8680737752062002,
"learning_rate": 2.9991820948514795e-06,
"loss": 0.4608,
"step": 614
},
{
"epoch": 0.7728709721758595,
"grad_norm": 0.9257236575469265,
"learning_rate": 2.9678750323848893e-06,
"loss": 0.4786,
"step": 615
},
{
"epoch": 0.7741276729436252,
"grad_norm": 0.9656830199257891,
"learning_rate": 2.936703730454017e-06,
"loss": 0.4358,
"step": 616
},
{
"epoch": 0.7753843737113908,
"grad_norm": 0.9907979170488217,
"learning_rate": 2.90566879084461e-06,
"loss": 0.5215,
"step": 617
},
{
"epoch": 0.7766410744791564,
"grad_norm": 0.8950799317069913,
"learning_rate": 2.8747708127098593e-06,
"loss": 0.4717,
"step": 618
},
{
"epoch": 0.777897775246922,
"grad_norm": 0.9505878825468562,
"learning_rate": 2.8440103925587904e-06,
"loss": 0.4566,
"step": 619
},
{
"epoch": 0.7791544760146877,
"grad_norm": 0.969358314749328,
"learning_rate": 2.813388124244778e-06,
"loss": 0.4492,
"step": 620
},
{
"epoch": 0.7804111767824533,
"grad_norm": 0.9785261324929386,
"learning_rate": 2.7829045989540594e-06,
"loss": 0.4496,
"step": 621
},
{
"epoch": 0.7816678775502189,
"grad_norm": 0.9531236242065506,
"learning_rate": 2.7525604051943512e-06,
"loss": 0.5504,
"step": 622
},
{
"epoch": 0.7829245783179846,
"grad_norm": 0.9826298336215402,
"learning_rate": 2.7223561287834467e-06,
"loss": 0.4586,
"step": 623
},
{
"epoch": 0.7841812790857502,
"grad_norm": 0.9317085993058872,
"learning_rate": 2.692292352837942e-06,
"loss": 0.4672,
"step": 624
},
{
"epoch": 0.7854379798535158,
"grad_norm": 1.0787276642963513,
"learning_rate": 2.662369657761963e-06,
"loss": 0.4608,
"step": 625
},
{
"epoch": 0.7866946806212815,
"grad_norm": 0.9372652077711696,
"learning_rate": 2.6325886212359496e-06,
"loss": 0.4377,
"step": 626
},
{
"epoch": 0.7879513813890471,
"grad_norm": 0.9921327711014636,
"learning_rate": 2.602949818205539e-06,
"loss": 0.4584,
"step": 627
},
{
"epoch": 0.7892080821568127,
"grad_norm": 0.947977203839485,
"learning_rate": 2.5734538208704197e-06,
"loss": 0.4787,
"step": 628
},
{
"epoch": 0.7904647829245783,
"grad_norm": 0.8656699847159132,
"learning_rate": 2.5441011986733165e-06,
"loss": 0.4662,
"step": 629
},
{
"epoch": 0.791721483692344,
"grad_norm": 0.8459597720147866,
"learning_rate": 2.514892518288988e-06,
"loss": 0.4475,
"step": 630
},
{
"epoch": 0.7929781844601096,
"grad_norm": 0.9446896479383777,
"learning_rate": 2.485828343613288e-06,
"loss": 0.4766,
"step": 631
},
{
"epoch": 0.7942348852278752,
"grad_norm": 0.9122011845063069,
"learning_rate": 2.456909235752276e-06,
"loss": 0.4502,
"step": 632
},
{
"epoch": 0.7954915859956408,
"grad_norm": 0.9193582934450675,
"learning_rate": 2.4281357530113804e-06,
"loss": 0.4823,
"step": 633
},
{
"epoch": 0.7967482867634065,
"grad_norm": 0.897269277270419,
"learning_rate": 2.399508450884631e-06,
"loss": 0.4513,
"step": 634
},
{
"epoch": 0.7980049875311721,
"grad_norm": 0.9475888045070303,
"learning_rate": 2.3710278820439313e-06,
"loss": 0.5161,
"step": 635
},
{
"epoch": 0.7992616882989377,
"grad_norm": 0.9443257123202231,
"learning_rate": 2.3426945963283853e-06,
"loss": 0.486,
"step": 636
},
{
"epoch": 0.8005183890667034,
"grad_norm": 0.9502264882194557,
"learning_rate": 2.3145091407336785e-06,
"loss": 0.4578,
"step": 637
},
{
"epoch": 0.801775089834469,
"grad_norm": 0.9187557661433506,
"learning_rate": 2.2864720594015288e-06,
"loss": 0.4871,
"step": 638
},
{
"epoch": 0.8030317906022346,
"grad_norm": 0.8862615742754856,
"learning_rate": 2.2585838936091753e-06,
"loss": 0.4635,
"step": 639
},
{
"epoch": 0.8042884913700002,
"grad_norm": 0.9044746570210405,
"learning_rate": 2.230845181758928e-06,
"loss": 0.4614,
"step": 640
},
{
"epoch": 0.8055451921377659,
"grad_norm": 0.9371401414038523,
"learning_rate": 2.2032564593677773e-06,
"loss": 0.4718,
"step": 641
},
{
"epoch": 0.8068018929055314,
"grad_norm": 1.0205326351085469,
"learning_rate": 2.1758182590570454e-06,
"loss": 0.4587,
"step": 642
},
{
"epoch": 0.808058593673297,
"grad_norm": 0.9391567214691986,
"learning_rate": 2.148531110542118e-06,
"loss": 0.4583,
"step": 643
},
{
"epoch": 0.8093152944410626,
"grad_norm": 0.9423357541683243,
"learning_rate": 2.1213955406222076e-06,
"loss": 0.4564,
"step": 644
},
{
"epoch": 0.8105719952088283,
"grad_norm": 0.9170267994353932,
"learning_rate": 2.09441207317019e-06,
"loss": 0.4501,
"step": 645
},
{
"epoch": 0.8118286959765939,
"grad_norm": 0.9670344205389225,
"learning_rate": 2.0675812291224796e-06,
"loss": 0.4827,
"step": 646
},
{
"epoch": 0.8130853967443595,
"grad_norm": 0.9406467714039654,
"learning_rate": 2.0409035264689857e-06,
"loss": 0.4708,
"step": 647
},
{
"epoch": 0.8143420975121252,
"grad_norm": 0.8958190576761381,
"learning_rate": 2.014379480243105e-06,
"loss": 0.4483,
"step": 648
},
{
"epoch": 0.8155987982798908,
"grad_norm": 0.9488884622008084,
"learning_rate": 1.988009602511779e-06,
"loss": 0.4569,
"step": 649
},
{
"epoch": 0.8168554990476564,
"grad_norm": 0.9215947064515085,
"learning_rate": 1.961794402365611e-06,
"loss": 0.5286,
"step": 650
},
{
"epoch": 0.8168554990476564,
"eval_loss": NaN,
"eval_runtime": 392.1209,
"eval_samples_per_second": 21.205,
"eval_steps_per_second": 2.652,
"step": 650
},
{
"epoch": 0.818112199815422,
"grad_norm": 0.9209169808681984,
"learning_rate": 1.935734385909028e-06,
"loss": 0.4795,
"step": 651
},
{
"epoch": 0.8193689005831877,
"grad_norm": 0.9076583048514367,
"learning_rate": 1.9098300562505266e-06,
"loss": 0.4711,
"step": 652
},
{
"epoch": 0.8206256013509533,
"grad_norm": 0.9912038417618557,
"learning_rate": 1.8840819134929467e-06,
"loss": 0.48,
"step": 653
},
{
"epoch": 0.8218823021187189,
"grad_norm": 0.9324565329660811,
"learning_rate": 1.8584904547238214e-06,
"loss": 0.4387,
"step": 654
},
{
"epoch": 0.8231390028864846,
"grad_norm": 0.8864965286423587,
"learning_rate": 1.8330561740057839e-06,
"loss": 0.4469,
"step": 655
},
{
"epoch": 0.8243957036542502,
"grad_norm": 0.8720461851671827,
"learning_rate": 1.8077795623670135e-06,
"loss": 0.4531,
"step": 656
},
{
"epoch": 0.8256524044220158,
"grad_norm": 0.8921315350992695,
"learning_rate": 1.7826611077917843e-06,
"loss": 0.4797,
"step": 657
},
{
"epoch": 0.8269091051897814,
"grad_norm": 0.8844082344199783,
"learning_rate": 1.757701295211014e-06,
"loss": 0.4523,
"step": 658
},
{
"epoch": 0.8281658059575471,
"grad_norm": 0.9020082915341568,
"learning_rate": 1.7329006064929232e-06,
"loss": 0.4772,
"step": 659
},
{
"epoch": 0.8294225067253127,
"grad_norm": 0.9058399917219788,
"learning_rate": 1.7082595204337183e-06,
"loss": 0.5386,
"step": 660
},
{
"epoch": 0.8306792074930783,
"grad_norm": 0.8817763873468203,
"learning_rate": 1.683778512748362e-06,
"loss": 0.4663,
"step": 661
},
{
"epoch": 0.831935908260844,
"grad_norm": 0.9634025095210257,
"learning_rate": 1.6594580560613782e-06,
"loss": 0.4689,
"step": 662
},
{
"epoch": 0.8331926090286096,
"grad_norm": 0.9633653565356368,
"learning_rate": 1.6352986198977327e-06,
"loss": 0.4667,
"step": 663
},
{
"epoch": 0.8344493097963752,
"grad_norm": 0.9487086679856576,
"learning_rate": 1.6113006706737667e-06,
"loss": 0.5184,
"step": 664
},
{
"epoch": 0.8357060105641408,
"grad_norm": 0.9325953615840198,
"learning_rate": 1.587464671688187e-06,
"loss": 0.456,
"step": 665
},
{
"epoch": 0.8369627113319065,
"grad_norm": 1.009772295164291,
"learning_rate": 1.563791083113142e-06,
"loss": 0.4451,
"step": 666
},
{
"epoch": 0.8382194120996721,
"grad_norm": 0.9419871710371429,
"learning_rate": 1.540280361985308e-06,
"loss": 0.5302,
"step": 667
},
{
"epoch": 0.8394761128674377,
"grad_norm": 0.8754091668537753,
"learning_rate": 1.5169329621970918e-06,
"loss": 0.4655,
"step": 668
},
{
"epoch": 0.8407328136352034,
"grad_norm": 0.8613931367520573,
"learning_rate": 1.4937493344878474e-06,
"loss": 0.63,
"step": 669
},
{
"epoch": 0.841989514402969,
"grad_norm": 0.9312238100708667,
"learning_rate": 1.4707299264351914e-06,
"loss": 0.5655,
"step": 670
},
{
"epoch": 0.8432462151707346,
"grad_norm": 0.9049299425860378,
"learning_rate": 1.4478751824463543e-06,
"loss": 0.4651,
"step": 671
},
{
"epoch": 0.8445029159385002,
"grad_norm": 0.8788628461504958,
"learning_rate": 1.4251855437495976e-06,
"loss": 0.475,
"step": 672
},
{
"epoch": 0.8457596167062659,
"grad_norm": 0.8951500164340301,
"learning_rate": 1.4026614483857037e-06,
"loss": 0.4673,
"step": 673
},
{
"epoch": 0.8470163174740315,
"grad_norm": 0.9244238966775989,
"learning_rate": 1.3803033311995072e-06,
"loss": 0.4796,
"step": 674
},
{
"epoch": 0.8482730182417971,
"grad_norm": 0.9749389580422814,
"learning_rate": 1.3581116238315194e-06,
"loss": 0.4735,
"step": 675
},
{
"epoch": 0.8495297190095628,
"grad_norm": 0.936241513202334,
"learning_rate": 1.336086754709569e-06,
"loss": 0.4635,
"step": 676
},
{
"epoch": 0.8507864197773284,
"grad_norm": 0.9584973374609863,
"learning_rate": 1.3142291490405568e-06,
"loss": 0.4638,
"step": 677
},
{
"epoch": 0.852043120545094,
"grad_norm": 0.9235602347304942,
"learning_rate": 1.2925392288022299e-06,
"loss": 0.4699,
"step": 678
},
{
"epoch": 0.8532998213128595,
"grad_norm": 0.8876268215636179,
"learning_rate": 1.2710174127350362e-06,
"loss": 0.4681,
"step": 679
},
{
"epoch": 0.8545565220806252,
"grad_norm": 0.9354140918132684,
"learning_rate": 1.2496641163340562e-06,
"loss": 0.4699,
"step": 680
},
{
"epoch": 0.8558132228483908,
"grad_norm": 0.9255534990449884,
"learning_rate": 1.2284797518409575e-06,
"loss": 0.4483,
"step": 681
},
{
"epoch": 0.8570699236161564,
"grad_norm": 0.9733945197204105,
"learning_rate": 1.2074647282360573e-06,
"loss": 0.4697,
"step": 682
},
{
"epoch": 0.858326624383922,
"grad_norm": 0.9031964521291095,
"learning_rate": 1.1866194512304075e-06,
"loss": 0.4631,
"step": 683
},
{
"epoch": 0.8595833251516877,
"grad_norm": 0.8891842214060443,
"learning_rate": 1.165944323257986e-06,
"loss": 0.5275,
"step": 684
},
{
"epoch": 0.8608400259194533,
"grad_norm": 0.9289352194308863,
"learning_rate": 1.1454397434679022e-06,
"loss": 0.4573,
"step": 685
},
{
"epoch": 0.8620967266872189,
"grad_norm": 0.9246693948234718,
"learning_rate": 1.125106107716708e-06,
"loss": 0.4606,
"step": 686
},
{
"epoch": 0.8633534274549846,
"grad_norm": 0.8763495664310694,
"learning_rate": 1.10494380856075e-06,
"loss": 0.4315,
"step": 687
},
{
"epoch": 0.8646101282227502,
"grad_norm": 0.9677586559163779,
"learning_rate": 1.0849532352485903e-06,
"loss": 0.4771,
"step": 688
},
{
"epoch": 0.8658668289905158,
"grad_norm": 0.8612887488120702,
"learning_rate": 1.0651347737134965e-06,
"loss": 0.4706,
"step": 689
},
{
"epoch": 0.8671235297582814,
"grad_norm": 0.8842124304281852,
"learning_rate": 1.0454888065659775e-06,
"loss": 0.4548,
"step": 690
},
{
"epoch": 0.8683802305260471,
"grad_norm": 0.9277463249641559,
"learning_rate": 1.0260157130864178e-06,
"loss": 0.4567,
"step": 691
},
{
"epoch": 0.8696369312938127,
"grad_norm": 0.8437870865194692,
"learning_rate": 1.0067158692177325e-06,
"loss": 0.4597,
"step": 692
},
{
"epoch": 0.8708936320615783,
"grad_norm": 0.8643510843583194,
"learning_rate": 9.87589647558135e-07,
"loss": 0.4256,
"step": 693
},
{
"epoch": 0.872150332829344,
"grad_norm": 0.9295154059632877,
"learning_rate": 9.686374173539147e-07,
"loss": 0.4861,
"step": 694
},
{
"epoch": 0.8734070335971096,
"grad_norm": 0.9448636463228528,
"learning_rate": 9.49859544492332e-07,
"loss": 0.4711,
"step": 695
},
{
"epoch": 0.8746637343648752,
"grad_norm": 0.9093250850860881,
"learning_rate": 9.312563914945461e-07,
"loss": 0.4696,
"step": 696
},
{
"epoch": 0.8759204351326408,
"grad_norm": 0.9002792388799274,
"learning_rate": 9.128283175086106e-07,
"loss": 0.4656,
"step": 697
},
{
"epoch": 0.8771771359004065,
"grad_norm": 0.8559481875479032,
"learning_rate": 8.945756783025528e-07,
"loss": 0.4335,
"step": 698
},
{
"epoch": 0.8784338366681721,
"grad_norm": 0.8796349450465737,
"learning_rate": 8.76498826257488e-07,
"loss": 0.4561,
"step": 699
},
{
"epoch": 0.8796905374359377,
"grad_norm": 0.8726954083378391,
"learning_rate": 8.585981103608343e-07,
"loss": 0.4993,
"step": 700
},
{
"epoch": 0.8796905374359377,
"eval_loss": NaN,
"eval_runtime": 391.2376,
"eval_samples_per_second": 21.253,
"eval_steps_per_second": 2.658,
"step": 700
},
{
"epoch": 0.8809472382037034,
"grad_norm": 1.4368360063954035,
"learning_rate": 8.40873876199565e-07,
"loss": 0.4788,
"step": 701
},
{
"epoch": 0.882203938971469,
"grad_norm": 0.9404327931258755,
"learning_rate": 8.233264659535367e-07,
"loss": 0.4502,
"step": 702
},
{
"epoch": 0.8834606397392346,
"grad_norm": 0.8946470160723485,
"learning_rate": 8.059562183888903e-07,
"loss": 0.4608,
"step": 703
},
{
"epoch": 0.8847173405070002,
"grad_norm": 0.9041054772329139,
"learning_rate": 7.887634688515e-07,
"loss": 0.4506,
"step": 704
},
{
"epoch": 0.8859740412747659,
"grad_norm": 0.8915805709724289,
"learning_rate": 7.71748549260507e-07,
"loss": 0.4424,
"step": 705
},
{
"epoch": 0.8872307420425315,
"grad_norm": 0.9307904793059376,
"learning_rate": 7.549117881019141e-07,
"loss": 0.4603,
"step": 706
},
{
"epoch": 0.8884874428102971,
"grad_norm": 0.9166282118331757,
"learning_rate": 7.382535104222366e-07,
"loss": 0.4669,
"step": 707
},
{
"epoch": 0.8897441435780628,
"grad_norm": 0.9726682302204435,
"learning_rate": 7.21774037822226e-07,
"loss": 0.4575,
"step": 708
},
{
"epoch": 0.8910008443458284,
"grad_norm": 0.9024738036393012,
"learning_rate": 7.054736884506718e-07,
"loss": 0.4584,
"step": 709
},
{
"epoch": 0.892257545113594,
"grad_norm": 0.9450717020332904,
"learning_rate": 6.8935277699825e-07,
"loss": 0.4478,
"step": 710
},
{
"epoch": 0.8935142458813596,
"grad_norm": 0.8734170696346301,
"learning_rate": 6.734116146914516e-07,
"loss": 0.4811,
"step": 711
},
{
"epoch": 0.8947709466491253,
"grad_norm": 0.9009909790129396,
"learning_rate": 6.576505092865748e-07,
"loss": 0.4695,
"step": 712
},
{
"epoch": 0.8960276474168909,
"grad_norm": 0.9465203253593445,
"learning_rate": 6.420697650637753e-07,
"loss": 0.4469,
"step": 713
},
{
"epoch": 0.8972843481846565,
"grad_norm": 0.9084056997490763,
"learning_rate": 6.266696828212071e-07,
"loss": 0.4467,
"step": 714
},
{
"epoch": 0.898541048952422,
"grad_norm": 0.9510351940546182,
"learning_rate": 6.114505598692011e-07,
"loss": 0.4538,
"step": 715
},
{
"epoch": 0.8997977497201877,
"grad_norm": 0.8690165913406015,
"learning_rate": 5.964126900245359e-07,
"loss": 0.4542,
"step": 716
},
{
"epoch": 0.9010544504879533,
"grad_norm": 0.8525566457571607,
"learning_rate": 5.815563636047539e-07,
"loss": 0.4526,
"step": 717
},
{
"epoch": 0.9023111512557189,
"grad_norm": 0.9287838751774835,
"learning_rate": 5.668818674225684e-07,
"loss": 0.4444,
"step": 718
},
{
"epoch": 0.9035678520234846,
"grad_norm": 0.960597993996023,
"learning_rate": 5.523894847803235e-07,
"loss": 0.4688,
"step": 719
},
{
"epoch": 0.9048245527912502,
"grad_norm": 0.8867015416832417,
"learning_rate": 5.380794954645141e-07,
"loss": 0.4369,
"step": 720
},
{
"epoch": 0.9060812535590158,
"grad_norm": 0.9250364567041895,
"learning_rate": 5.23952175740402e-07,
"loss": 0.4728,
"step": 721
},
{
"epoch": 0.9073379543267814,
"grad_norm": 0.8538358598265039,
"learning_rate": 5.100077983466667e-07,
"loss": 0.454,
"step": 722
},
{
"epoch": 0.9085946550945471,
"grad_norm": 0.8588765866461067,
"learning_rate": 4.962466324901483e-07,
"loss": 0.4488,
"step": 723
},
{
"epoch": 0.9098513558623127,
"grad_norm": 0.8879641890521957,
"learning_rate": 4.826689438406495e-07,
"loss": 0.4538,
"step": 724
},
{
"epoch": 0.9111080566300783,
"grad_norm": 0.868954790800622,
"learning_rate": 4.6927499452580574e-07,
"loss": 0.4258,
"step": 725
},
{
"epoch": 0.912364757397844,
"grad_norm": 0.869194697723244,
"learning_rate": 4.5606504312602384e-07,
"loss": 0.4428,
"step": 726
},
{
"epoch": 0.9136214581656096,
"grad_norm": 0.9689587398952784,
"learning_rate": 4.4303934466948804e-07,
"loss": 0.4461,
"step": 727
},
{
"epoch": 0.9148781589333752,
"grad_norm": 0.9714671146836645,
"learning_rate": 4.3019815062724567e-07,
"loss": 0.4523,
"step": 728
},
{
"epoch": 0.9161348597011408,
"grad_norm": 0.9098783404381591,
"learning_rate": 4.1754170890833777e-07,
"loss": 0.5114,
"step": 729
},
{
"epoch": 0.9173915604689065,
"grad_norm": 0.903763846717378,
"learning_rate": 4.0507026385502747e-07,
"loss": 0.4558,
"step": 730
},
{
"epoch": 0.9186482612366721,
"grad_norm": 0.8589002511680877,
"learning_rate": 3.9278405623806914e-07,
"loss": 0.4463,
"step": 731
},
{
"epoch": 0.9199049620044377,
"grad_norm": 0.9149570489401863,
"learning_rate": 3.806833232520746e-07,
"loss": 0.4665,
"step": 732
},
{
"epoch": 0.9211616627722033,
"grad_norm": 0.857139001990913,
"learning_rate": 3.687682985109209e-07,
"loss": 0.4847,
"step": 733
},
{
"epoch": 0.922418363539969,
"grad_norm": 0.9093058792535011,
"learning_rate": 3.5703921204324863e-07,
"loss": 0.5461,
"step": 734
},
{
"epoch": 0.9236750643077346,
"grad_norm": 0.8979631581554675,
"learning_rate": 3.454962902880199e-07,
"loss": 0.4589,
"step": 735
},
{
"epoch": 0.9249317650755002,
"grad_norm": 0.8828698198732804,
"learning_rate": 3.3413975609013713e-07,
"loss": 0.4622,
"step": 736
},
{
"epoch": 0.9261884658432659,
"grad_norm": 0.8945978464715129,
"learning_rate": 3.2296982869616134e-07,
"loss": 0.4247,
"step": 737
},
{
"epoch": 0.9274451666110315,
"grad_norm": 0.8897356645887511,
"learning_rate": 3.1198672375005403e-07,
"loss": 0.465,
"step": 738
},
{
"epoch": 0.9287018673787971,
"grad_norm": 0.8698017477642928,
"learning_rate": 3.0119065328903517e-07,
"loss": 0.4426,
"step": 739
},
{
"epoch": 0.9299585681465627,
"grad_norm": 0.8856161042535935,
"learning_rate": 2.905818257394799e-07,
"loss": 0.4489,
"step": 740
},
{
"epoch": 0.9312152689143284,
"grad_norm": 0.8551132648021337,
"learning_rate": 2.801604459128926e-07,
"loss": 0.4546,
"step": 741
},
{
"epoch": 0.932471969682094,
"grad_norm": 0.8755685457795177,
"learning_rate": 2.6992671500196134e-07,
"loss": 0.4767,
"step": 742
},
{
"epoch": 0.9337286704498596,
"grad_norm": 1.0076154176090533,
"learning_rate": 2.5988083057666534e-07,
"loss": 0.5176,
"step": 743
},
{
"epoch": 0.9349853712176253,
"grad_norm": 0.8773948389601972,
"learning_rate": 2.5002298658046484e-07,
"loss": 0.4665,
"step": 744
},
{
"epoch": 0.9362420719853909,
"grad_norm": 0.949607330031333,
"learning_rate": 2.4035337332655504e-07,
"loss": 0.4865,
"step": 745
},
{
"epoch": 0.9374987727531565,
"grad_norm": 0.8727367626946688,
"learning_rate": 2.308721774941991e-07,
"loss": 0.4755,
"step": 746
},
{
"epoch": 0.9387554735209221,
"grad_norm": 0.950145974990549,
"learning_rate": 2.2157958212510877e-07,
"loss": 0.4658,
"step": 747
},
{
"epoch": 0.9400121742886878,
"grad_norm": 0.8949527330544057,
"learning_rate": 2.124757666199273e-07,
"loss": 0.4715,
"step": 748
},
{
"epoch": 0.9412688750564534,
"grad_norm": 0.857039311943978,
"learning_rate": 2.035609067347566e-07,
"loss": 0.4846,
"step": 749
},
{
"epoch": 0.942525575824219,
"grad_norm": 0.9246764854504824,
"learning_rate": 1.9483517457776436e-07,
"loss": 0.5387,
"step": 750
},
{
"epoch": 0.942525575824219,
"eval_loss": NaN,
"eval_runtime": 392.0537,
"eval_samples_per_second": 21.209,
"eval_steps_per_second": 2.653,
"step": 750
},
{
"epoch": 0.9437822765919847,
"grad_norm": 0.9246699749677241,
"learning_rate": 1.8629873860586567e-07,
"loss": 0.4915,
"step": 751
},
{
"epoch": 0.9450389773597502,
"grad_norm": 0.9273150008131176,
"learning_rate": 1.7795176362146783e-07,
"loss": 0.4394,
"step": 752
},
{
"epoch": 0.9462956781275158,
"grad_norm": 0.8721007270250941,
"learning_rate": 1.6979441076928837e-07,
"loss": 0.5125,
"step": 753
},
{
"epoch": 0.9475523788952814,
"grad_norm": 0.8233991769635042,
"learning_rate": 1.6182683753324435e-07,
"loss": 0.4622,
"step": 754
},
{
"epoch": 0.9488090796630471,
"grad_norm": 0.88887926972337,
"learning_rate": 1.5404919773341576e-07,
"loss": 0.46,
"step": 755
},
{
"epoch": 0.9500657804308127,
"grad_norm": 0.9151783514520999,
"learning_rate": 1.464616415230702e-07,
"loss": 0.4769,
"step": 756
},
{
"epoch": 0.9513224811985783,
"grad_norm": 0.9233338402724911,
"learning_rate": 1.3906431538576626e-07,
"loss": 0.4511,
"step": 757
},
{
"epoch": 0.952579181966344,
"grad_norm": 0.876720746013767,
"learning_rate": 1.3185736213252808e-07,
"loss": 0.4623,
"step": 758
},
{
"epoch": 0.9538358827341096,
"grad_norm": 0.8808110259800377,
"learning_rate": 1.2484092089908307e-07,
"loss": 0.455,
"step": 759
},
{
"epoch": 0.9550925835018752,
"grad_norm": 0.9127364361814908,
"learning_rate": 1.1801512714318286e-07,
"loss": 0.4659,
"step": 760
},
{
"epoch": 0.9563492842696408,
"grad_norm": 0.9250866501406244,
"learning_rate": 1.113801126419789e-07,
"loss": 0.4097,
"step": 761
},
{
"epoch": 0.9576059850374065,
"grad_norm": 0.9039383706969123,
"learning_rate": 1.0493600548948879e-07,
"loss": 0.4372,
"step": 762
},
{
"epoch": 0.9588626858051721,
"grad_norm": 0.9023237019743435,
"learning_rate": 9.8682930094115e-08,
"loss": 0.4524,
"step": 763
},
{
"epoch": 0.9601193865729377,
"grad_norm": 0.9011103800480463,
"learning_rate": 9.262100717624678e-08,
"loss": 0.4342,
"step": 764
},
{
"epoch": 0.9613760873407033,
"grad_norm": 0.9408044434789575,
"learning_rate": 8.675035376593088e-08,
"loss": 0.4699,
"step": 765
},
{
"epoch": 0.962632788108469,
"grad_norm": 0.8157071127061982,
"learning_rate": 8.107108320060675e-08,
"loss": 0.4388,
"step": 766
},
{
"epoch": 0.9638894888762346,
"grad_norm": 0.9614204495709833,
"learning_rate": 7.558330512292378e-08,
"loss": 0.5011,
"step": 767
},
{
"epoch": 0.9651461896440002,
"grad_norm": 0.9164749653087414,
"learning_rate": 7.028712547862526e-08,
"loss": 0.4922,
"step": 768
},
{
"epoch": 0.9664028904117659,
"grad_norm": 0.8941273006248021,
"learning_rate": 6.51826465144978e-08,
"loss": 0.5093,
"step": 769
},
{
"epoch": 0.9676595911795315,
"grad_norm": 0.9345588962411722,
"learning_rate": 6.026996677640062e-08,
"loss": 0.4654,
"step": 770
},
{
"epoch": 0.9689162919472971,
"grad_norm": 0.8714785559856224,
"learning_rate": 5.5549181107362734e-08,
"loss": 0.4459,
"step": 771
},
{
"epoch": 0.9701729927150627,
"grad_norm": 0.8661748110904719,
"learning_rate": 5.102038064575099e-08,
"loss": 0.426,
"step": 772
},
{
"epoch": 0.9714296934828284,
"grad_norm": 0.8998862817234337,
"learning_rate": 4.6683652823513725e-08,
"loss": 0.4808,
"step": 773
},
{
"epoch": 0.972686394250594,
"grad_norm": 0.9253080744252631,
"learning_rate": 4.253908136448881e-08,
"loss": 0.4777,
"step": 774
},
{
"epoch": 0.9739430950183596,
"grad_norm": 0.8884730525355212,
"learning_rate": 3.858674628278825e-08,
"loss": 0.4445,
"step": 775
},
{
"epoch": 0.9751997957861253,
"grad_norm": 0.8920492529443076,
"learning_rate": 3.482672388125719e-08,
"loss": 0.5067,
"step": 776
},
{
"epoch": 0.9764564965538909,
"grad_norm": 0.913098536113785,
"learning_rate": 3.125908674999289e-08,
"loss": 0.4704,
"step": 777
},
{
"epoch": 0.9777131973216565,
"grad_norm": 0.8585074814418806,
"learning_rate": 2.7883903764953647e-08,
"loss": 0.4383,
"step": 778
},
{
"epoch": 0.9789698980894221,
"grad_norm": 0.8563259132871073,
"learning_rate": 2.470124008661978e-08,
"loss": 0.486,
"step": 779
},
{
"epoch": 0.9802265988571878,
"grad_norm": 1.0110036318343043,
"learning_rate": 2.171115715874139e-08,
"loss": 0.4784,
"step": 780
},
{
"epoch": 0.9814832996249534,
"grad_norm": 0.9352424779414596,
"learning_rate": 1.8913712707149255e-08,
"loss": 0.4732,
"step": 781
},
{
"epoch": 0.982740000392719,
"grad_norm": 0.904141538390942,
"learning_rate": 1.630896073864352e-08,
"loss": 0.483,
"step": 782
},
{
"epoch": 0.9839967011604847,
"grad_norm": 0.8991415167948376,
"learning_rate": 1.3896951539945635e-08,
"loss": 0.467,
"step": 783
},
{
"epoch": 0.9852534019282503,
"grad_norm": 0.8724840641479233,
"learning_rate": 1.1677731676733584e-08,
"loss": 0.4614,
"step": 784
},
{
"epoch": 0.9865101026960159,
"grad_norm": 0.8862179253481299,
"learning_rate": 9.651343992740369e-09,
"loss": 0.4452,
"step": 785
},
{
"epoch": 0.9877668034637815,
"grad_norm": 0.877231477260857,
"learning_rate": 7.817827608924689e-09,
"loss": 0.4913,
"step": 786
},
{
"epoch": 0.9890235042315472,
"grad_norm": 0.9256633248853225,
"learning_rate": 6.1772179227181926e-09,
"loss": 0.4854,
"step": 787
},
{
"epoch": 0.9902802049993128,
"grad_norm": 0.8460307202589556,
"learning_rate": 4.7295466073427055e-09,
"loss": 0.4562,
"step": 788
},
{
"epoch": 0.9915369057670783,
"grad_norm": 0.9076188371683147,
"learning_rate": 3.474841611197377e-09,
"loss": 0.4736,
"step": 789
},
{
"epoch": 0.9927936065348439,
"grad_norm": 0.8533095890208698,
"learning_rate": 2.4131271573191172e-09,
"loss": 0.452,
"step": 790
},
{
"epoch": 0.9940503073026096,
"grad_norm": 0.8331866413387022,
"learning_rate": 1.5444237429140806e-09,
"loss": 0.4387,
"step": 791
},
{
"epoch": 0.9953070080703752,
"grad_norm": 0.8888758833913929,
"learning_rate": 8.687481389657582e-10,
"loss": 0.4634,
"step": 792
},
{
"epoch": 0.9965637088381408,
"grad_norm": 0.9295696536473508,
"learning_rate": 3.861133899063507e-10,
"loss": 0.4674,
"step": 793
},
{
"epoch": 0.9978204096059065,
"grad_norm": 0.9051031551266895,
"learning_rate": 9.652881336696951e-11,
"loss": 0.4903,
"step": 794
},
{
"epoch": 0.9990771103736721,
"grad_norm": 0.9088993808171332,
"learning_rate": 0.0,
"loss": 0.4562,
"step": 795
}
],
"logging_steps": 1.0,
"max_steps": 795,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1331628437667840.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}