chansung's picture
Model save
b26371f verified
raw
history blame
No virus
37.5 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 1045,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0009569377990430622,
"grad_norm": 948.5667114257812,
"learning_rate": 2.8571428571428573e-06,
"loss": 46.4048,
"step": 1
},
{
"epoch": 0.004784688995215311,
"grad_norm": 579.7166137695312,
"learning_rate": 1.4285714285714284e-05,
"loss": 42.2625,
"step": 5
},
{
"epoch": 0.009569377990430622,
"grad_norm": 96.71955871582031,
"learning_rate": 2.8571428571428567e-05,
"loss": 24.8784,
"step": 10
},
{
"epoch": 0.014354066985645933,
"grad_norm": 14.768364906311035,
"learning_rate": 4.285714285714285e-05,
"loss": 20.4002,
"step": 15
},
{
"epoch": 0.019138755980861243,
"grad_norm": 12.786796569824219,
"learning_rate": 5.7142857142857135e-05,
"loss": 18.6753,
"step": 20
},
{
"epoch": 0.023923444976076555,
"grad_norm": 46.72217559814453,
"learning_rate": 7.142857142857142e-05,
"loss": 15.9881,
"step": 25
},
{
"epoch": 0.028708133971291867,
"grad_norm": 16.096284866333008,
"learning_rate": 8.57142857142857e-05,
"loss": 7.4899,
"step": 30
},
{
"epoch": 0.03349282296650718,
"grad_norm": 6.774743556976318,
"learning_rate": 9.999999999999999e-05,
"loss": 2.28,
"step": 35
},
{
"epoch": 0.03827751196172249,
"grad_norm": 25.32003402709961,
"learning_rate": 0.00011428571428571427,
"loss": 1.816,
"step": 40
},
{
"epoch": 0.0430622009569378,
"grad_norm": 4.912317276000977,
"learning_rate": 0.00012857142857142855,
"loss": 1.6463,
"step": 45
},
{
"epoch": 0.04784688995215311,
"grad_norm": 3.605388879776001,
"learning_rate": 0.00014285714285714284,
"loss": 1.5122,
"step": 50
},
{
"epoch": 0.05263157894736842,
"grad_norm": 5.295993328094482,
"learning_rate": 0.00015714285714285713,
"loss": 1.419,
"step": 55
},
{
"epoch": 0.05741626794258373,
"grad_norm": 9.990426063537598,
"learning_rate": 0.0001714285714285714,
"loss": 1.3221,
"step": 60
},
{
"epoch": 0.06220095693779904,
"grad_norm": 28.818571090698242,
"learning_rate": 0.00018571428571428572,
"loss": 1.2352,
"step": 65
},
{
"epoch": 0.06698564593301436,
"grad_norm": 5.248025417327881,
"learning_rate": 0.00019999999999999998,
"loss": 1.2013,
"step": 70
},
{
"epoch": 0.07177033492822966,
"grad_norm": 4.098865509033203,
"learning_rate": 0.00021428571428571427,
"loss": 1.1641,
"step": 75
},
{
"epoch": 0.07655502392344497,
"grad_norm": 1.8723022937774658,
"learning_rate": 0.00022857142857142854,
"loss": 1.1404,
"step": 80
},
{
"epoch": 0.08133971291866028,
"grad_norm": 20.416040420532227,
"learning_rate": 0.00024285714285714283,
"loss": 1.1758,
"step": 85
},
{
"epoch": 0.0861244019138756,
"grad_norm": 2.548046112060547,
"learning_rate": 0.0002571428571428571,
"loss": 1.1864,
"step": 90
},
{
"epoch": 0.09090909090909091,
"grad_norm": 10.036776542663574,
"learning_rate": 0.0002714285714285714,
"loss": 1.1193,
"step": 95
},
{
"epoch": 0.09569377990430622,
"grad_norm": 7.473161697387695,
"learning_rate": 0.0002857142857142857,
"loss": 1.1753,
"step": 100
},
{
"epoch": 0.10047846889952153,
"grad_norm": 1.7201554775238037,
"learning_rate": 0.0003,
"loss": 1.1511,
"step": 105
},
{
"epoch": 0.10526315789473684,
"grad_norm": 22.742843627929688,
"learning_rate": 0.00029997905717787856,
"loss": 1.1194,
"step": 110
},
{
"epoch": 0.11004784688995216,
"grad_norm": 4.5689191818237305,
"learning_rate": 0.00029991623455953814,
"loss": 1.1401,
"step": 115
},
{
"epoch": 0.11483253588516747,
"grad_norm": 4.430986404418945,
"learning_rate": 0.00029981154968741785,
"loss": 1.1312,
"step": 120
},
{
"epoch": 0.11961722488038277,
"grad_norm": 9.391127586364746,
"learning_rate": 0.000299665031793473,
"loss": 1.0676,
"step": 125
},
{
"epoch": 0.12440191387559808,
"grad_norm": 10.183479309082031,
"learning_rate": 0.0002994767217910127,
"loss": 1.1184,
"step": 130
},
{
"epoch": 0.1291866028708134,
"grad_norm": 1.511924147605896,
"learning_rate": 0.00029924667226327557,
"loss": 1.0999,
"step": 135
},
{
"epoch": 0.1339712918660287,
"grad_norm": 8.494267463684082,
"learning_rate": 0.0002989749474487461,
"loss": 1.0913,
"step": 140
},
{
"epoch": 0.13875598086124402,
"grad_norm": 2.4349634647369385,
"learning_rate": 0.000298661623223217,
"loss": 1.0802,
"step": 145
},
{
"epoch": 0.14354066985645933,
"grad_norm": 1.914871096611023,
"learning_rate": 0.0002983067870786019,
"loss": 1.0589,
"step": 150
},
{
"epoch": 0.14832535885167464,
"grad_norm": 8.084524154663086,
"learning_rate": 0.00029791053809850426,
"loss": 1.0661,
"step": 155
},
{
"epoch": 0.15311004784688995,
"grad_norm": 15.332684516906738,
"learning_rate": 0.0002974729869305495,
"loss": 1.0688,
"step": 160
},
{
"epoch": 0.15789473684210525,
"grad_norm": 30.3629207611084,
"learning_rate": 0.000296994255755488,
"loss": 1.0531,
"step": 165
},
{
"epoch": 0.16267942583732056,
"grad_norm": 6.318498611450195,
"learning_rate": 0.0002964744782530777,
"loss": 1.0664,
"step": 170
},
{
"epoch": 0.1674641148325359,
"grad_norm": 7.628419399261475,
"learning_rate": 0.0002959137995647556,
"loss": 1.1012,
"step": 175
},
{
"epoch": 0.1722488038277512,
"grad_norm": 6.398703098297119,
"learning_rate": 0.0002953123762531088,
"loss": 1.1012,
"step": 180
},
{
"epoch": 0.17703349282296652,
"grad_norm": 1.5758147239685059,
"learning_rate": 0.00029467037625815644,
"loss": 1.0655,
"step": 185
},
{
"epoch": 0.18181818181818182,
"grad_norm": 3.354144811630249,
"learning_rate": 0.0002939879788504546,
"loss": 1.0405,
"step": 190
},
{
"epoch": 0.18660287081339713,
"grad_norm": 7.777205467224121,
"learning_rate": 0.00029326537458103683,
"loss": 1.0575,
"step": 195
},
{
"epoch": 0.19138755980861244,
"grad_norm": 3.233966588973999,
"learning_rate": 0.0002925027652282056,
"loss": 1.0519,
"step": 200
},
{
"epoch": 0.19617224880382775,
"grad_norm": 5.466601848602295,
"learning_rate": 0.00029170036374118777,
"loss": 1.0377,
"step": 205
},
{
"epoch": 0.20095693779904306,
"grad_norm": 2.6383893489837646,
"learning_rate": 0.0002908583941806717,
"loss": 1.0435,
"step": 210
},
{
"epoch": 0.20574162679425836,
"grad_norm": 1.6238977909088135,
"learning_rate": 0.00028997709165624034,
"loss": 1.0098,
"step": 215
},
{
"epoch": 0.21052631578947367,
"grad_norm": 4.616154193878174,
"learning_rate": 0.0002890567022607206,
"loss": 1.007,
"step": 220
},
{
"epoch": 0.215311004784689,
"grad_norm": 3.3014893531799316,
"learning_rate": 0.0002880974830014643,
"loss": 1.0311,
"step": 225
},
{
"epoch": 0.22009569377990432,
"grad_norm": 1.7747435569763184,
"learning_rate": 0.0002870997017285824,
"loss": 0.9898,
"step": 230
},
{
"epoch": 0.22488038277511962,
"grad_norm": 2.502115488052368,
"learning_rate": 0.0002860636370601511,
"loss": 1.0031,
"step": 235
},
{
"epoch": 0.22966507177033493,
"grad_norm": 2.33097767829895,
"learning_rate": 0.00028498957830441117,
"loss": 1.0175,
"step": 240
},
{
"epoch": 0.23444976076555024,
"grad_norm": 2.667079448699951,
"learning_rate": 0.00028387782537898215,
"loss": 1.0102,
"step": 245
},
{
"epoch": 0.23923444976076555,
"grad_norm": 5.493870735168457,
"learning_rate": 0.0002827286887271143,
"loss": 0.9815,
"step": 250
},
{
"epoch": 0.24401913875598086,
"grad_norm": 1.950860619544983,
"learning_rate": 0.0002815424892310007,
"loss": 0.9865,
"step": 255
},
{
"epoch": 0.24880382775119617,
"grad_norm": 1.8500947952270508,
"learning_rate": 0.00028031955812217544,
"loss": 0.9819,
"step": 260
},
{
"epoch": 0.2535885167464115,
"grad_norm": 4.860908508300781,
"learning_rate": 0.0002790602368890209,
"loss": 0.9728,
"step": 265
},
{
"epoch": 0.2583732057416268,
"grad_norm": 2.0686490535736084,
"learning_rate": 0.0002777648771814114,
"loss": 1.02,
"step": 270
},
{
"epoch": 0.2631578947368421,
"grad_norm": 3.445263624191284,
"learning_rate": 0.00027643384071251954,
"loss": 1.0279,
"step": 275
},
{
"epoch": 0.2679425837320574,
"grad_norm": 1.4115147590637207,
"learning_rate": 0.000275067499157812,
"loss": 0.9813,
"step": 280
},
{
"epoch": 0.2727272727272727,
"grad_norm": 1.2224127054214478,
"learning_rate": 0.00027366623405126404,
"loss": 0.9903,
"step": 285
},
{
"epoch": 0.27751196172248804,
"grad_norm": 1.8986269235610962,
"learning_rate": 0.0002722304366788205,
"loss": 1.0094,
"step": 290
},
{
"epoch": 0.2822966507177033,
"grad_norm": 4.729682445526123,
"learning_rate": 0.0002707605079691344,
"loss": 1.0046,
"step": 295
},
{
"epoch": 0.28708133971291866,
"grad_norm": 2.3022851943969727,
"learning_rate": 0.0002692568583816124,
"loss": 1.0305,
"step": 300
},
{
"epoch": 0.291866028708134,
"grad_norm": 2.677311658859253,
"learning_rate": 0.0002677199077917991,
"loss": 1.0308,
"step": 305
},
{
"epoch": 0.2966507177033493,
"grad_norm": 1.3505252599716187,
"learning_rate": 0.000266150085374132,
"loss": 0.9911,
"step": 310
},
{
"epoch": 0.3014354066985646,
"grad_norm": 1.311340093612671,
"learning_rate": 0.00026454782948209983,
"loss": 0.9897,
"step": 315
},
{
"epoch": 0.3062200956937799,
"grad_norm": 1.3043484687805176,
"learning_rate": 0.00026291358752583764,
"loss": 0.9659,
"step": 320
},
{
"epoch": 0.31100478468899523,
"grad_norm": 1.1022688150405884,
"learning_rate": 0.0002612478158471936,
"loss": 0.9852,
"step": 325
},
{
"epoch": 0.3157894736842105,
"grad_norm": 1.8880759477615356,
"learning_rate": 0.0002595509795923004,
"loss": 0.9707,
"step": 330
},
{
"epoch": 0.32057416267942584,
"grad_norm": 4.946929454803467,
"learning_rate": 0.0002578235525816894,
"loss": 1.0037,
"step": 335
},
{
"epoch": 0.3253588516746411,
"grad_norm": 1.5874648094177246,
"learning_rate": 0.00025606601717798207,
"loss": 1.013,
"step": 340
},
{
"epoch": 0.33014354066985646,
"grad_norm": 1.1188175678253174,
"learning_rate": 0.0002542788641511963,
"loss": 0.9814,
"step": 345
},
{
"epoch": 0.3349282296650718,
"grad_norm": 1.3674702644348145,
"learning_rate": 0.0002524625925417046,
"loss": 0.9924,
"step": 350
},
{
"epoch": 0.3397129186602871,
"grad_norm": 1.3748054504394531,
"learning_rate": 0.0002506177095208835,
"loss": 0.9952,
"step": 355
},
{
"epoch": 0.3444976076555024,
"grad_norm": 1.4951032400131226,
"learning_rate": 0.00024874473024949224,
"loss": 0.9859,
"step": 360
},
{
"epoch": 0.3492822966507177,
"grad_norm": 1.5338605642318726,
"learning_rate": 0.0002468441777338203,
"loss": 0.9986,
"step": 365
},
{
"epoch": 0.35406698564593303,
"grad_norm": 2.292693614959717,
"learning_rate": 0.00024491658267964474,
"loss": 0.9815,
"step": 370
},
{
"epoch": 0.3588516746411483,
"grad_norm": 3.6766607761383057,
"learning_rate": 0.00024296248334403672,
"loss": 0.9696,
"step": 375
},
{
"epoch": 0.36363636363636365,
"grad_norm": 1.8351837396621704,
"learning_rate": 0.00024098242538506007,
"loss": 0.9806,
"step": 380
},
{
"epoch": 0.3684210526315789,
"grad_norm": 3.7850887775421143,
"learning_rate": 0.00023897696170940326,
"loss": 0.9922,
"step": 385
},
{
"epoch": 0.37320574162679426,
"grad_norm": 1.6278636455535889,
"learning_rate": 0.0002369466523179866,
"loss": 0.9524,
"step": 390
},
{
"epoch": 0.37799043062200954,
"grad_norm": 2.634073495864868,
"learning_rate": 0.0002348920641495893,
"loss": 0.9721,
"step": 395
},
{
"epoch": 0.3827751196172249,
"grad_norm": 1.0519402027130127,
"learning_rate": 0.0002328137709225385,
"loss": 0.9433,
"step": 400
},
{
"epoch": 0.3875598086124402,
"grad_norm": 1.6269092559814453,
"learning_rate": 0.00023071235297450588,
"loss": 0.965,
"step": 405
},
{
"epoch": 0.3923444976076555,
"grad_norm": 1.0608420372009277,
"learning_rate": 0.0002285883971004553,
"loss": 0.9458,
"step": 410
},
{
"epoch": 0.39712918660287083,
"grad_norm": 1.1753120422363281,
"learning_rate": 0.00022644249638878762,
"loss": 0.9708,
"step": 415
},
{
"epoch": 0.4019138755980861,
"grad_norm": 1.1925132274627686,
"learning_rate": 0.000224275250055728,
"loss": 0.9419,
"step": 420
},
{
"epoch": 0.40669856459330145,
"grad_norm": 2.7282474040985107,
"learning_rate": 0.00022208726327800255,
"loss": 0.9776,
"step": 425
},
{
"epoch": 0.41148325358851673,
"grad_norm": 1.2128546237945557,
"learning_rate": 0.0002198791470238497,
"loss": 0.9726,
"step": 430
},
{
"epoch": 0.41626794258373206,
"grad_norm": 1.216680884361267,
"learning_rate": 0.00021765151788241525,
"loss": 0.965,
"step": 435
},
{
"epoch": 0.42105263157894735,
"grad_norm": 1.2095612287521362,
"learning_rate": 0.0002154049978915774,
"loss": 0.9734,
"step": 440
},
{
"epoch": 0.4258373205741627,
"grad_norm": 2.2863378524780273,
"learning_rate": 0.00021314021436425024,
"loss": 0.9668,
"step": 445
},
{
"epoch": 0.430622009569378,
"grad_norm": 1.6475805044174194,
"learning_rate": 0.00021085779971321454,
"loss": 0.953,
"step": 450
},
{
"epoch": 0.4354066985645933,
"grad_norm": 2.581031560897827,
"learning_rate": 0.0002085583912745242,
"loss": 0.9481,
"step": 455
},
{
"epoch": 0.44019138755980863,
"grad_norm": 1.6886945962905884,
"learning_rate": 0.00020624263112953815,
"loss": 0.9521,
"step": 460
},
{
"epoch": 0.4449760765550239,
"grad_norm": 1.6784495115280151,
"learning_rate": 0.0002039111659256269,
"loss": 0.9448,
"step": 465
},
{
"epoch": 0.44976076555023925,
"grad_norm": 1.741603970527649,
"learning_rate": 0.00020156464669560448,
"loss": 0.9396,
"step": 470
},
{
"epoch": 0.45454545454545453,
"grad_norm": 1.4577537775039673,
"learning_rate": 0.00019920372867593533,
"loss": 0.9381,
"step": 475
},
{
"epoch": 0.45933014354066987,
"grad_norm": 1.2143363952636719,
"learning_rate": 0.00019682907112376796,
"loss": 0.942,
"step": 480
},
{
"epoch": 0.46411483253588515,
"grad_norm": 1.133268117904663,
"learning_rate": 0.0001944413371328451,
"loss": 0.9307,
"step": 485
},
{
"epoch": 0.4688995215311005,
"grad_norm": 1.3096511363983154,
"learning_rate": 0.0001920411934483434,
"loss": 0.9489,
"step": 490
},
{
"epoch": 0.47368421052631576,
"grad_norm": 1.1326684951782227,
"learning_rate": 0.00018962931028069293,
"loss": 0.9481,
"step": 495
},
{
"epoch": 0.4784688995215311,
"grad_norm": 1.435521125793457,
"learning_rate": 0.00018720636111842898,
"loss": 0.941,
"step": 500
},
{
"epoch": 0.48325358851674644,
"grad_norm": 1.5249536037445068,
"learning_rate": 0.00018477302254012924,
"loss": 0.9466,
"step": 505
},
{
"epoch": 0.4880382775119617,
"grad_norm": 1.5078822374343872,
"learning_rate": 0.000182329974025487,
"loss": 0.9407,
"step": 510
},
{
"epoch": 0.49282296650717705,
"grad_norm": 1.1270455121994019,
"learning_rate": 0.0001798778977655754,
"loss": 0.9342,
"step": 515
},
{
"epoch": 0.49760765550239233,
"grad_norm": 1.8892922401428223,
"learning_rate": 0.00017741747847235353,
"loss": 0.9215,
"step": 520
},
{
"epoch": 0.5023923444976076,
"grad_norm": 1.1376285552978516,
"learning_rate": 0.0001749494031874695,
"loss": 0.941,
"step": 525
},
{
"epoch": 0.507177033492823,
"grad_norm": 1.3463551998138428,
"learning_rate": 0.0001724743610904122,
"loss": 0.9494,
"step": 530
},
{
"epoch": 0.5119617224880383,
"grad_norm": 2.9610071182250977,
"learning_rate": 0.00016999304330606683,
"loss": 0.9342,
"step": 535
},
{
"epoch": 0.5167464114832536,
"grad_norm": 1.1553057432174683,
"learning_rate": 0.00016750614271172643,
"loss": 0.9421,
"step": 540
},
{
"epoch": 0.5215311004784688,
"grad_norm": 1.025996446609497,
"learning_rate": 0.00016501435374361475,
"loss": 0.9165,
"step": 545
},
{
"epoch": 0.5263157894736842,
"grad_norm": 1.1247053146362305,
"learning_rate": 0.00016251837220297347,
"loss": 0.9382,
"step": 550
},
{
"epoch": 0.5311004784688995,
"grad_norm": 1.027742624282837,
"learning_rate": 0.00016001889506176824,
"loss": 0.9461,
"step": 555
},
{
"epoch": 0.5358851674641149,
"grad_norm": 1.2423522472381592,
"learning_rate": 0.00015751662026806788,
"loss": 0.9411,
"step": 560
},
{
"epoch": 0.5406698564593302,
"grad_norm": 1.5147805213928223,
"learning_rate": 0.00015501224655115118,
"loss": 0.9334,
"step": 565
},
{
"epoch": 0.5454545454545454,
"grad_norm": 1.2580386400222778,
"learning_rate": 0.00015250647322639512,
"loss": 0.9258,
"step": 570
},
{
"epoch": 0.5502392344497608,
"grad_norm": 1.1148898601531982,
"learning_rate": 0.00015,
"loss": 0.92,
"step": 575
},
{
"epoch": 0.5550239234449761,
"grad_norm": 1.6557159423828125,
"learning_rate": 0.00014749352677360483,
"loss": 0.9124,
"step": 580
},
{
"epoch": 0.5598086124401914,
"grad_norm": 1.3142054080963135,
"learning_rate": 0.00014498775344884884,
"loss": 0.9292,
"step": 585
},
{
"epoch": 0.5645933014354066,
"grad_norm": 1.3109759092330933,
"learning_rate": 0.00014248337973193212,
"loss": 0.9075,
"step": 590
},
{
"epoch": 0.569377990430622,
"grad_norm": 1.8519434928894043,
"learning_rate": 0.00013998110493823176,
"loss": 0.9334,
"step": 595
},
{
"epoch": 0.5741626794258373,
"grad_norm": 1.467532992362976,
"learning_rate": 0.00013748162779702653,
"loss": 0.9038,
"step": 600
},
{
"epoch": 0.5789473684210527,
"grad_norm": 1.2827484607696533,
"learning_rate": 0.00013498564625638522,
"loss": 0.9209,
"step": 605
},
{
"epoch": 0.583732057416268,
"grad_norm": 2.0187387466430664,
"learning_rate": 0.00013249385728827357,
"loss": 0.909,
"step": 610
},
{
"epoch": 0.5885167464114832,
"grad_norm": 1.0019465684890747,
"learning_rate": 0.0001300069566939332,
"loss": 0.9324,
"step": 615
},
{
"epoch": 0.5933014354066986,
"grad_norm": 1.1702934503555298,
"learning_rate": 0.00012752563890958778,
"loss": 0.8818,
"step": 620
},
{
"epoch": 0.5980861244019139,
"grad_norm": 1.1036059856414795,
"learning_rate": 0.0001250505968125305,
"loss": 0.9082,
"step": 625
},
{
"epoch": 0.6028708133971292,
"grad_norm": 1.3649473190307617,
"learning_rate": 0.00012258252152764647,
"loss": 0.9077,
"step": 630
},
{
"epoch": 0.6076555023923444,
"grad_norm": 1.2819594144821167,
"learning_rate": 0.0001201221022344246,
"loss": 0.899,
"step": 635
},
{
"epoch": 0.6124401913875598,
"grad_norm": 1.1259446144104004,
"learning_rate": 0.00011767002597451296,
"loss": 0.8888,
"step": 640
},
{
"epoch": 0.6172248803827751,
"grad_norm": 1.0908528566360474,
"learning_rate": 0.00011522697745987075,
"loss": 0.8954,
"step": 645
},
{
"epoch": 0.6220095693779905,
"grad_norm": 1.0630327463150024,
"learning_rate": 0.000112793638881571,
"loss": 0.8974,
"step": 650
},
{
"epoch": 0.6267942583732058,
"grad_norm": 1.2224838733673096,
"learning_rate": 0.00011037068971930709,
"loss": 0.8945,
"step": 655
},
{
"epoch": 0.631578947368421,
"grad_norm": 1.2856760025024414,
"learning_rate": 0.00010795880655165656,
"loss": 0.8994,
"step": 660
},
{
"epoch": 0.6363636363636364,
"grad_norm": 1.0856343507766724,
"learning_rate": 0.0001055586628671549,
"loss": 0.8883,
"step": 665
},
{
"epoch": 0.6411483253588517,
"grad_norm": 1.063491940498352,
"learning_rate": 0.00010317092887623205,
"loss": 0.8866,
"step": 670
},
{
"epoch": 0.645933014354067,
"grad_norm": 1.2210334539413452,
"learning_rate": 0.00010079627132406462,
"loss": 0.8902,
"step": 675
},
{
"epoch": 0.6507177033492823,
"grad_norm": 0.9613766074180603,
"learning_rate": 9.843535330439555e-05,
"loss": 0.8748,
"step": 680
},
{
"epoch": 0.6555023923444976,
"grad_norm": 0.9286556839942932,
"learning_rate": 9.608883407437309e-05,
"loss": 0.8824,
"step": 685
},
{
"epoch": 0.6602870813397129,
"grad_norm": 1.13881254196167,
"learning_rate": 9.375736887046185e-05,
"loss": 0.8779,
"step": 690
},
{
"epoch": 0.6650717703349283,
"grad_norm": 1.128664255142212,
"learning_rate": 9.144160872547578e-05,
"loss": 0.8623,
"step": 695
},
{
"epoch": 0.6698564593301436,
"grad_norm": 1.3374972343444824,
"learning_rate": 8.914220028678546e-05,
"loss": 0.8831,
"step": 700
},
{
"epoch": 0.6746411483253588,
"grad_norm": 1.5959270000457764,
"learning_rate": 8.685978563574976e-05,
"loss": 0.8956,
"step": 705
},
{
"epoch": 0.6794258373205742,
"grad_norm": 1.582465410232544,
"learning_rate": 8.459500210842261e-05,
"loss": 0.8804,
"step": 710
},
{
"epoch": 0.6842105263157895,
"grad_norm": 1.3153728246688843,
"learning_rate": 8.23484821175847e-05,
"loss": 0.8776,
"step": 715
},
{
"epoch": 0.6889952153110048,
"grad_norm": 1.1769953966140747,
"learning_rate": 8.012085297615027e-05,
"loss": 0.8777,
"step": 720
},
{
"epoch": 0.69377990430622,
"grad_norm": 1.1776398420333862,
"learning_rate": 7.791273672199742e-05,
"loss": 0.8866,
"step": 725
},
{
"epoch": 0.6985645933014354,
"grad_norm": 1.0515143871307373,
"learning_rate": 7.572474994427197e-05,
"loss": 0.8887,
"step": 730
},
{
"epoch": 0.7033492822966507,
"grad_norm": 1.0670396089553833,
"learning_rate": 7.355750361121237e-05,
"loss": 0.8631,
"step": 735
},
{
"epoch": 0.7081339712918661,
"grad_norm": 1.3679234981536865,
"learning_rate": 7.141160289954471e-05,
"loss": 0.8701,
"step": 740
},
{
"epoch": 0.7129186602870813,
"grad_norm": 1.0791184902191162,
"learning_rate": 6.92876470254941e-05,
"loss": 0.879,
"step": 745
},
{
"epoch": 0.7177033492822966,
"grad_norm": 1.0610390901565552,
"learning_rate": 6.718622907746151e-05,
"loss": 0.8617,
"step": 750
},
{
"epoch": 0.722488038277512,
"grad_norm": 1.062545657157898,
"learning_rate": 6.51079358504107e-05,
"loss": 0.8743,
"step": 755
},
{
"epoch": 0.7272727272727273,
"grad_norm": 0.8820713758468628,
"learning_rate": 6.305334768201336e-05,
"loss": 0.8553,
"step": 760
},
{
"epoch": 0.7320574162679426,
"grad_norm": 0.9072192311286926,
"learning_rate": 6.1023038290596715e-05,
"loss": 0.8613,
"step": 765
},
{
"epoch": 0.7368421052631579,
"grad_norm": 1.0919402837753296,
"learning_rate": 5.901757461493988e-05,
"loss": 0.8679,
"step": 770
},
{
"epoch": 0.7416267942583732,
"grad_norm": 1.0981807708740234,
"learning_rate": 5.703751665596328e-05,
"loss": 0.8665,
"step": 775
},
{
"epoch": 0.7464114832535885,
"grad_norm": 1.0239495038986206,
"learning_rate": 5.50834173203552e-05,
"loss": 0.8819,
"step": 780
},
{
"epoch": 0.7511961722488039,
"grad_norm": 1.0707833766937256,
"learning_rate": 5.315582226617963e-05,
"loss": 0.8571,
"step": 785
},
{
"epoch": 0.7559808612440191,
"grad_norm": 1.0176221132278442,
"learning_rate": 5.125526975050776e-05,
"loss": 0.8718,
"step": 790
},
{
"epoch": 0.7607655502392344,
"grad_norm": 0.9414092898368835,
"learning_rate": 4.938229047911651e-05,
"loss": 0.8547,
"step": 795
},
{
"epoch": 0.7655502392344498,
"grad_norm": 0.8967716097831726,
"learning_rate": 4.753740745829537e-05,
"loss": 0.8385,
"step": 800
},
{
"epoch": 0.7703349282296651,
"grad_norm": 1.160251498222351,
"learning_rate": 4.5721135848803653e-05,
"loss": 0.8471,
"step": 805
},
{
"epoch": 0.7751196172248804,
"grad_norm": 0.9440346360206604,
"learning_rate": 4.3933982822017876e-05,
"loss": 0.8526,
"step": 810
},
{
"epoch": 0.7799043062200957,
"grad_norm": 0.9394163489341736,
"learning_rate": 4.217644741831062e-05,
"loss": 0.8661,
"step": 815
},
{
"epoch": 0.784688995215311,
"grad_norm": 0.9960381388664246,
"learning_rate": 4.044902040769963e-05,
"loss": 0.8848,
"step": 820
},
{
"epoch": 0.7894736842105263,
"grad_norm": 0.9392317533493042,
"learning_rate": 3.875218415280636e-05,
"loss": 0.8562,
"step": 825
},
{
"epoch": 0.7942583732057417,
"grad_norm": 1.0181559324264526,
"learning_rate": 3.708641247416229e-05,
"loss": 0.854,
"step": 830
},
{
"epoch": 0.7990430622009569,
"grad_norm": 1.1440843343734741,
"learning_rate": 3.54521705179002e-05,
"loss": 0.8597,
"step": 835
},
{
"epoch": 0.8038277511961722,
"grad_norm": 1.0670745372772217,
"learning_rate": 3.3849914625868004e-05,
"loss": 0.8847,
"step": 840
},
{
"epoch": 0.8086124401913876,
"grad_norm": 0.9774266481399536,
"learning_rate": 3.228009220820085e-05,
"loss": 0.8673,
"step": 845
},
{
"epoch": 0.8133971291866029,
"grad_norm": 0.9691421389579773,
"learning_rate": 3.074314161838754e-05,
"loss": 0.8315,
"step": 850
},
{
"epoch": 0.8181818181818182,
"grad_norm": 0.9607205986976624,
"learning_rate": 2.923949203086557e-05,
"loss": 0.862,
"step": 855
},
{
"epoch": 0.8229665071770335,
"grad_norm": 1.0411772727966309,
"learning_rate": 2.77695633211795e-05,
"loss": 0.8659,
"step": 860
},
{
"epoch": 0.8277511961722488,
"grad_norm": 1.090226173400879,
"learning_rate": 2.6333765948735986e-05,
"loss": 0.8491,
"step": 865
},
{
"epoch": 0.8325358851674641,
"grad_norm": 0.8712325096130371,
"learning_rate": 2.493250084218795e-05,
"loss": 0.8626,
"step": 870
},
{
"epoch": 0.8373205741626795,
"grad_norm": 0.8832561373710632,
"learning_rate": 2.3566159287480447e-05,
"loss": 0.8479,
"step": 875
},
{
"epoch": 0.8421052631578947,
"grad_norm": 0.9606965184211731,
"learning_rate": 2.2235122818588565e-05,
"loss": 0.8625,
"step": 880
},
{
"epoch": 0.84688995215311,
"grad_norm": 1.056420087814331,
"learning_rate": 2.0939763110979125e-05,
"loss": 0.863,
"step": 885
},
{
"epoch": 0.8516746411483254,
"grad_norm": 0.9405905604362488,
"learning_rate": 1.9680441877824527e-05,
"loss": 0.853,
"step": 890
},
{
"epoch": 0.8564593301435407,
"grad_norm": 0.8905633687973022,
"learning_rate": 1.8457510768999273e-05,
"loss": 0.8431,
"step": 895
},
{
"epoch": 0.861244019138756,
"grad_norm": 0.981576144695282,
"learning_rate": 1.727131127288572e-05,
"loss": 0.8477,
"step": 900
},
{
"epoch": 0.8660287081339713,
"grad_norm": 1.0129060745239258,
"learning_rate": 1.612217462101783e-05,
"loss": 0.8384,
"step": 905
},
{
"epoch": 0.8708133971291866,
"grad_norm": 0.9811695218086243,
"learning_rate": 1.5010421695588825e-05,
"loss": 0.8573,
"step": 910
},
{
"epoch": 0.8755980861244019,
"grad_norm": 0.9269863963127136,
"learning_rate": 1.3936362939848893e-05,
"loss": 0.8561,
"step": 915
},
{
"epoch": 0.8803827751196173,
"grad_norm": 0.8773042559623718,
"learning_rate": 1.2900298271417592e-05,
"loss": 0.8203,
"step": 920
},
{
"epoch": 0.8851674641148325,
"grad_norm": 0.9804287552833557,
"learning_rate": 1.1902516998535666e-05,
"loss": 0.8382,
"step": 925
},
{
"epoch": 0.8899521531100478,
"grad_norm": 0.8934711217880249,
"learning_rate": 1.0943297739279395e-05,
"loss": 0.8416,
"step": 930
},
{
"epoch": 0.8947368421052632,
"grad_norm": 1.0209717750549316,
"learning_rate": 1.0022908343759616e-05,
"loss": 0.8595,
"step": 935
},
{
"epoch": 0.8995215311004785,
"grad_norm": 0.8881580829620361,
"learning_rate": 9.141605819328318e-06,
"loss": 0.8332,
"step": 940
},
{
"epoch": 0.9043062200956937,
"grad_norm": 0.8664154410362244,
"learning_rate": 8.299636258812197e-06,
"loss": 0.8437,
"step": 945
},
{
"epoch": 0.9090909090909091,
"grad_norm": 0.8912498950958252,
"learning_rate": 7.4972347717943916e-06,
"loss": 0.8485,
"step": 950
},
{
"epoch": 0.9138755980861244,
"grad_norm": 0.9088477492332458,
"learning_rate": 6.734625418963135e-06,
"loss": 0.8472,
"step": 955
},
{
"epoch": 0.9186602870813397,
"grad_norm": 0.9696795344352722,
"learning_rate": 6.012021149545382e-06,
"loss": 0.8548,
"step": 960
},
{
"epoch": 0.9234449760765551,
"grad_norm": 0.9158973097801208,
"learning_rate": 5.329623741843531e-06,
"loss": 0.8333,
"step": 965
},
{
"epoch": 0.9282296650717703,
"grad_norm": 0.9025459289550781,
"learning_rate": 4.6876237468912e-06,
"loss": 0.8422,
"step": 970
},
{
"epoch": 0.9330143540669856,
"grad_norm": 0.8703954815864563,
"learning_rate": 4.086200435244441e-06,
"loss": 0.8265,
"step": 975
},
{
"epoch": 0.937799043062201,
"grad_norm": 0.9033318161964417,
"learning_rate": 3.5255217469222942e-06,
"loss": 0.8203,
"step": 980
},
{
"epoch": 0.9425837320574163,
"grad_norm": 1.0247541666030884,
"learning_rate": 3.0057442445119872e-06,
"loss": 0.8296,
"step": 985
},
{
"epoch": 0.9473684210526315,
"grad_norm": 0.9569628238677979,
"learning_rate": 2.5270130694504733e-06,
"loss": 0.8644,
"step": 990
},
{
"epoch": 0.9521531100478469,
"grad_norm": 0.8948321342468262,
"learning_rate": 2.089461901495715e-06,
"loss": 0.8278,
"step": 995
},
{
"epoch": 0.9569377990430622,
"grad_norm": 0.969680666923523,
"learning_rate": 1.6932129213980793e-06,
"loss": 0.8452,
"step": 1000
},
{
"epoch": 0.9617224880382775,
"grad_norm": 0.9193556904792786,
"learning_rate": 1.3383767767829956e-06,
"loss": 0.8341,
"step": 1005
},
{
"epoch": 0.9665071770334929,
"grad_norm": 0.9120749235153198,
"learning_rate": 1.0250525512538855e-06,
"loss": 0.8359,
"step": 1010
},
{
"epoch": 0.9712918660287081,
"grad_norm": 0.8678485155105591,
"learning_rate": 7.533277367243795e-07,
"loss": 0.8187,
"step": 1015
},
{
"epoch": 0.9760765550239234,
"grad_norm": 1.054465889930725,
"learning_rate": 5.232782089872601e-07,
"loss": 0.8487,
"step": 1020
},
{
"epoch": 0.9808612440191388,
"grad_norm": 0.8861187696456909,
"learning_rate": 3.349682065270254e-07,
"loss": 0.8504,
"step": 1025
},
{
"epoch": 0.9856459330143541,
"grad_norm": 0.9375335574150085,
"learning_rate": 1.8845031258213905e-07,
"loss": 0.8459,
"step": 1030
},
{
"epoch": 0.9904306220095693,
"grad_norm": 0.9127489924430847,
"learning_rate": 8.376544046180822e-08,
"loss": 0.8447,
"step": 1035
},
{
"epoch": 0.9952153110047847,
"grad_norm": 0.8907783031463623,
"learning_rate": 2.0942822121439164e-08,
"loss": 0.8246,
"step": 1040
},
{
"epoch": 1.0,
"grad_norm": 0.9073578715324402,
"learning_rate": 0.0,
"loss": 0.8382,
"step": 1045
},
{
"epoch": 1.0,
"eval_loss": 2.4202542304992676,
"eval_runtime": 1.1602,
"eval_samples_per_second": 8.619,
"eval_steps_per_second": 0.862,
"step": 1045
},
{
"epoch": 1.0,
"step": 1045,
"total_flos": 6.049215074407547e+17,
"train_loss": 1.5597172723432478,
"train_runtime": 7161.753,
"train_samples_per_second": 1.751,
"train_steps_per_second": 0.146
}
],
"logging_steps": 5,
"max_steps": 1045,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 6.049215074407547e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}