{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.6913319238900635, "eval_steps": 500, "global_step": 200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.042283298097251586, "grad_norm": 1.2008394002914429, "learning_rate": 4.9975392245612254e-05, "loss": 4.1473, "num_input_tokens_seen": 68384, "step": 5 }, { "epoch": 0.08456659619450317, "grad_norm": 1.1565380096435547, "learning_rate": 4.9901617425775067e-05, "loss": 4.021, "num_input_tokens_seen": 128224, "step": 10 }, { "epoch": 0.12684989429175475, "grad_norm": 0.972896158695221, "learning_rate": 4.9778820775100185e-05, "loss": 3.8212, "num_input_tokens_seen": 197760, "step": 15 }, { "epoch": 0.16913319238900634, "grad_norm": 1.0641047954559326, "learning_rate": 4.9607244033573156e-05, "loss": 3.7653, "num_input_tokens_seen": 262832, "step": 20 }, { "epoch": 0.21141649048625794, "grad_norm": 1.0994701385498047, "learning_rate": 4.93872249706591e-05, "loss": 3.7434, "num_input_tokens_seen": 333472, "step": 25 }, { "epoch": 0.2536997885835095, "grad_norm": 1.193864107131958, "learning_rate": 4.91191967203629e-05, "loss": 3.5488, "num_input_tokens_seen": 393616, "step": 30 }, { "epoch": 0.2959830866807611, "grad_norm": 1.0215297937393188, "learning_rate": 4.8803686928552736e-05, "loss": 3.5732, "num_input_tokens_seen": 458240, "step": 35 }, { "epoch": 0.3382663847780127, "grad_norm": 1.312198519706726, "learning_rate": 4.84413167142257e-05, "loss": 3.7108, "num_input_tokens_seen": 524576, "step": 40 }, { "epoch": 0.38054968287526425, "grad_norm": 1.3579617738723755, "learning_rate": 4.803279944676032e-05, "loss": 3.6871, "num_input_tokens_seen": 591856, "step": 45 }, { "epoch": 0.42283298097251587, "grad_norm": 1.4524191617965698, "learning_rate": 4.7578939341563095e-05, "loss": 3.286, "num_input_tokens_seen": 655648, "step": 50 }, { "epoch": 0.46511627906976744, "grad_norm": 1.248831033706665, "learning_rate": 4.70806298768736e-05, "loss": 3.5377, "num_input_tokens_seen": 721280, "step": 55 }, { "epoch": 0.507399577167019, "grad_norm": 1.2427473068237305, "learning_rate": 4.653885203484515e-05, "loss": 3.56, "num_input_tokens_seen": 784688, "step": 60 }, { "epoch": 0.5496828752642706, "grad_norm": 1.323653221130371, "learning_rate": 4.595467237036329e-05, "loss": 3.4937, "num_input_tokens_seen": 850848, "step": 65 }, { "epoch": 0.5919661733615222, "grad_norm": 1.3730229139328003, "learning_rate": 4.532924091140417e-05, "loss": 3.3823, "num_input_tokens_seen": 912480, "step": 70 }, { "epoch": 0.6342494714587738, "grad_norm": 1.5328903198242188, "learning_rate": 4.466378889506607e-05, "loss": 3.3798, "num_input_tokens_seen": 978448, "step": 75 }, { "epoch": 0.6765327695560254, "grad_norm": 1.4153543710708618, "learning_rate": 4.395962634373097e-05, "loss": 3.3044, "num_input_tokens_seen": 1041280, "step": 80 }, { "epoch": 0.718816067653277, "grad_norm": 1.6301600933074951, "learning_rate": 4.3218139486127854e-05, "loss": 3.3661, "num_input_tokens_seen": 1102224, "step": 85 }, { "epoch": 0.7610993657505285, "grad_norm": 1.6634522676467896, "learning_rate": 4.2440788028374624e-05, "loss": 3.3829, "num_input_tokens_seen": 1166576, "step": 90 }, { "epoch": 0.8033826638477801, "grad_norm": 1.4539167881011963, "learning_rate": 4.1629102280370904e-05, "loss": 3.2241, "num_input_tokens_seen": 1230096, "step": 95 }, { "epoch": 0.8456659619450317, "grad_norm": 1.453364372253418, "learning_rate": 4.0784680143198836e-05, "loss": 3.0931, "num_input_tokens_seen": 1297968, "step": 100 }, { "epoch": 0.8879492600422833, "grad_norm": 2.0559964179992676, "learning_rate": 3.990918396346254e-05, "loss": 3.3581, "num_input_tokens_seen": 1361760, "step": 105 }, { "epoch": 0.9302325581395349, "grad_norm": 2.0426218509674072, "learning_rate": 3.900433726075865e-05, "loss": 3.2308, "num_input_tokens_seen": 1422096, "step": 110 }, { "epoch": 0.9725158562367865, "grad_norm": 1.7696195840835571, "learning_rate": 3.8071921334720696e-05, "loss": 3.1427, "num_input_tokens_seen": 1491120, "step": 115 }, { "epoch": 1.014799154334038, "grad_norm": 2.3205981254577637, "learning_rate": 3.711377175831626e-05, "loss": 3.0988, "num_input_tokens_seen": 1556440, "step": 120 }, { "epoch": 1.0570824524312896, "grad_norm": 1.7631757259368896, "learning_rate": 3.613177476430079e-05, "loss": 3.0695, "num_input_tokens_seen": 1624200, "step": 125 }, { "epoch": 1.0993657505285412, "grad_norm": 2.32861328125, "learning_rate": 3.512786353194134e-05, "loss": 3.1109, "num_input_tokens_seen": 1686600, "step": 130 }, { "epoch": 1.1416490486257929, "grad_norm": 2.1291067600250244, "learning_rate": 3.410401438132056e-05, "loss": 3.116, "num_input_tokens_seen": 1755144, "step": 135 }, { "epoch": 1.1839323467230445, "grad_norm": 1.8158444166183472, "learning_rate": 3.3062242882712724e-05, "loss": 2.9528, "num_input_tokens_seen": 1819672, "step": 140 }, { "epoch": 1.226215644820296, "grad_norm": 2.161515951156616, "learning_rate": 3.200459988869111e-05, "loss": 2.8494, "num_input_tokens_seen": 1886136, "step": 145 }, { "epoch": 1.2684989429175475, "grad_norm": 2.321112632751465, "learning_rate": 3.093316749677788e-05, "loss": 3.1525, "num_input_tokens_seen": 1947656, "step": 150 }, { "epoch": 1.3107822410147991, "grad_norm": 2.4329445362091064, "learning_rate": 2.985005495058446e-05, "loss": 2.8991, "num_input_tokens_seen": 2013272, "step": 155 }, { "epoch": 1.3530655391120507, "grad_norm": 2.7929720878601074, "learning_rate": 2.875739448751176e-05, "loss": 3.026, "num_input_tokens_seen": 2079816, "step": 160 }, { "epoch": 1.3953488372093024, "grad_norm": 2.6059908866882324, "learning_rate": 2.7657337141184138e-05, "loss": 2.8813, "num_input_tokens_seen": 2146008, "step": 165 }, { "epoch": 1.437632135306554, "grad_norm": 2.831233501434326, "learning_rate": 2.655204850688085e-05, "loss": 2.7721, "num_input_tokens_seen": 2212840, "step": 170 }, { "epoch": 1.4799154334038054, "grad_norm": 2.757953643798828, "learning_rate": 2.5443704478301154e-05, "loss": 2.737, "num_input_tokens_seen": 2278824, "step": 175 }, { "epoch": 1.522198731501057, "grad_norm": 2.8655178546905518, "learning_rate": 2.433448696405563e-05, "loss": 2.6645, "num_input_tokens_seen": 2339768, "step": 180 }, { "epoch": 1.5644820295983086, "grad_norm": 2.8346216678619385, "learning_rate": 2.3226579592316538e-05, "loss": 2.7233, "num_input_tokens_seen": 2407976, "step": 185 }, { "epoch": 1.6067653276955602, "grad_norm": 2.432692766189575, "learning_rate": 2.2122163412082927e-05, "loss": 2.7463, "num_input_tokens_seen": 2469176, "step": 190 }, { "epoch": 1.6490486257928119, "grad_norm": 3.0815346240997314, "learning_rate": 2.1023412599523204e-05, "loss": 2.8348, "num_input_tokens_seen": 2528152, "step": 195 }, { "epoch": 1.6913319238900635, "grad_norm": 3.362581968307495, "learning_rate": 1.993249017784766e-05, "loss": 2.7732, "num_input_tokens_seen": 2588984, "step": 200 } ], "logging_steps": 5, "max_steps": 354, "num_input_tokens_seen": 2588984, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.083238062474035e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }