| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.9987174005985464, |
| "eval_steps": 500, |
| "global_step": 73, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.013681060282171868, |
| "grad_norm": 7.548368406170154, |
| "learning_rate": 0.0, |
| "loss": 1.5136, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.027362120564343735, |
| "grad_norm": 7.827536257097371, |
| "learning_rate": 6.666666666666667e-07, |
| "loss": 1.4973, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.041043180846515606, |
| "grad_norm": 7.398147035041813, |
| "learning_rate": 1.3333333333333334e-06, |
| "loss": 1.4526, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.05472424112868747, |
| "grad_norm": 6.621467563471504, |
| "learning_rate": 2.0000000000000003e-06, |
| "loss": 1.3913, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.06840530141085935, |
| "grad_norm": 5.273700898837342, |
| "learning_rate": 2.666666666666667e-06, |
| "loss": 1.3428, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.08208636169303121, |
| "grad_norm": 3.9736038530372646, |
| "learning_rate": 3.3333333333333333e-06, |
| "loss": 1.1501, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.09576742197520308, |
| "grad_norm": 3.5059552761650705, |
| "learning_rate": 4.000000000000001e-06, |
| "loss": 1.2428, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.10944848225737494, |
| "grad_norm": 3.6945698874338477, |
| "learning_rate": 4.666666666666667e-06, |
| "loss": 1.3626, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.12312954253954682, |
| "grad_norm": 2.9010276135864124, |
| "learning_rate": 5.333333333333334e-06, |
| "loss": 1.0364, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.1368106028217187, |
| "grad_norm": 3.6634274177297486, |
| "learning_rate": 6e-06, |
| "loss": 1.1989, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.15049166310389056, |
| "grad_norm": 3.3162811107208, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": 1.074, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.16417272338606242, |
| "grad_norm": 3.4273757924875965, |
| "learning_rate": 7.333333333333333e-06, |
| "loss": 0.9367, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.1778537836682343, |
| "grad_norm": 3.712520280275751, |
| "learning_rate": 8.000000000000001e-06, |
| "loss": 0.9526, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.19153484395040615, |
| "grad_norm": 3.489057750483782, |
| "learning_rate": 8.666666666666668e-06, |
| "loss": 1.124, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.20521590423257802, |
| "grad_norm": 2.588795249616481, |
| "learning_rate": 9.333333333333334e-06, |
| "loss": 0.8115, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.21889696451474988, |
| "grad_norm": 2.5083608844278915, |
| "learning_rate": 1e-05, |
| "loss": 0.8998, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.23257802479692177, |
| "grad_norm": 1.733597300225938, |
| "learning_rate": 9.998562273265786e-06, |
| "loss": 0.7702, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.24625908507909364, |
| "grad_norm": 1.91140737255229, |
| "learning_rate": 9.994249919886402e-06, |
| "loss": 0.8392, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.2599401453612655, |
| "grad_norm": 2.2610549095712646, |
| "learning_rate": 9.98706541985615e-06, |
| "loss": 0.9605, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.2736212056434374, |
| "grad_norm": 1.820469178389298, |
| "learning_rate": 9.977012904914133e-06, |
| "loss": 0.8495, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.28730226592560926, |
| "grad_norm": 1.8986542607595096, |
| "learning_rate": 9.964098156168143e-06, |
| "loss": 0.8255, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.3009833262077811, |
| "grad_norm": 1.8732046088429473, |
| "learning_rate": 9.948328600769996e-06, |
| "loss": 0.8314, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.314664386489953, |
| "grad_norm": 2.022587424257079, |
| "learning_rate": 9.929713307644245e-06, |
| "loss": 0.9394, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.32834544677212485, |
| "grad_norm": 2.185047520868034, |
| "learning_rate": 9.908262982272724e-06, |
| "loss": 0.9943, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.3420265070542967, |
| "grad_norm": 1.858153572034805, |
| "learning_rate": 9.883989960537934e-06, |
| "loss": 0.919, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.3557075673364686, |
| "grad_norm": 1.6533796703245207, |
| "learning_rate": 9.85690820162878e-06, |
| "loss": 0.9009, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.36938862761864044, |
| "grad_norm": 1.5392129172041236, |
| "learning_rate": 9.827033280012783e-06, |
| "loss": 0.8536, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.3830696879008123, |
| "grad_norm": 1.2934191579255045, |
| "learning_rate": 9.794382376479334e-06, |
| "loss": 0.7038, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.39675074818298417, |
| "grad_norm": 1.5922900609334563, |
| "learning_rate": 9.7589742682592e-06, |
| "loss": 0.8082, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.41043180846515603, |
| "grad_norm": 1.3349784246999898, |
| "learning_rate": 9.720829318225897e-06, |
| "loss": 0.7657, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.4241128687473279, |
| "grad_norm": 1.3968084241612435, |
| "learning_rate": 9.6799694631852e-06, |
| "loss": 0.7583, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.43779392902949976, |
| "grad_norm": 1.4438050306391899, |
| "learning_rate": 9.63641820125949e-06, |
| "loss": 0.8281, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.4514749893116717, |
| "grad_norm": 1.5653751821202764, |
| "learning_rate": 9.590200578374198e-06, |
| "loss": 0.8164, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.46515604959384355, |
| "grad_norm": 1.4428794862229695, |
| "learning_rate": 9.541343173854128e-06, |
| "loss": 0.7563, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.4788371098760154, |
| "grad_norm": 1.424184980221859, |
| "learning_rate": 9.48987408513794e-06, |
| "loss": 0.8012, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.4925181701581873, |
| "grad_norm": 1.4959423264612097, |
| "learning_rate": 9.435822911619564e-06, |
| "loss": 0.7733, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.5061992304403591, |
| "grad_norm": 1.5145765101418414, |
| "learning_rate": 9.379220737625877e-06, |
| "loss": 0.8185, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.519880290722531, |
| "grad_norm": 1.4205405113619893, |
| "learning_rate": 9.320100114540382e-06, |
| "loss": 0.7728, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.5335613510047028, |
| "grad_norm": 1.3618866165050287, |
| "learning_rate": 9.258495042083222e-06, |
| "loss": 0.7188, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.5472424112868748, |
| "grad_norm": 1.3210682200708475, |
| "learning_rate": 9.19444094875825e-06, |
| "loss": 0.7462, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.5609234715690466, |
| "grad_norm": 1.5209903146284076, |
| "learning_rate": 9.127974671478432e-06, |
| "loss": 0.8553, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.5746045318512185, |
| "grad_norm": 1.320169467583813, |
| "learning_rate": 9.059134434381274e-06, |
| "loss": 0.7292, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.5882855921333904, |
| "grad_norm": 1.3832163792888312, |
| "learning_rate": 8.987959826846479e-06, |
| "loss": 0.7342, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.6019666524155622, |
| "grad_norm": 1.3410523808663086, |
| "learning_rate": 8.914491780728471e-06, |
| "loss": 0.7135, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.6156477126977341, |
| "grad_norm": 1.7453570328171104, |
| "learning_rate": 8.838772546816857e-06, |
| "loss": 0.895, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.629328772979906, |
| "grad_norm": 1.3236234248181074, |
| "learning_rate": 8.760845670538387e-06, |
| "loss": 0.7094, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.6430098332620778, |
| "grad_norm": 1.3474642609149368, |
| "learning_rate": 8.6807559669144e-06, |
| "loss": 0.7463, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.6566908935442497, |
| "grad_norm": 1.3020601607506088, |
| "learning_rate": 8.598549494788111e-06, |
| "loss": 0.7731, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.6703719538264216, |
| "grad_norm": 1.5137293508647862, |
| "learning_rate": 8.5142735303366e-06, |
| "loss": 0.877, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.6840530141085934, |
| "grad_norm": 1.5403773102256209, |
| "learning_rate": 8.427976539882725e-06, |
| "loss": 0.8081, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.6977340743907653, |
| "grad_norm": 1.4219674838583494, |
| "learning_rate": 8.339708152022586e-06, |
| "loss": 0.7865, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.7114151346729372, |
| "grad_norm": 1.3635386118244723, |
| "learning_rate": 8.24951912908459e-06, |
| "loss": 0.7789, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.725096194955109, |
| "grad_norm": 1.2536175575645903, |
| "learning_rate": 8.157461337936506e-06, |
| "loss": 0.7167, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.7387772552372809, |
| "grad_norm": 1.399259576609449, |
| "learning_rate": 8.063587720157298e-06, |
| "loss": 0.7842, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.7524583155194527, |
| "grad_norm": 1.3669785990529744, |
| "learning_rate": 7.967952261590936e-06, |
| "loss": 0.8057, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.7661393758016246, |
| "grad_norm": 1.2116543523109478, |
| "learning_rate": 7.870609961299627e-06, |
| "loss": 0.71, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.7798204360837965, |
| "grad_norm": 1.496951541567372, |
| "learning_rate": 7.771616799934372e-06, |
| "loss": 0.9227, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.7935014963659683, |
| "grad_norm": 1.4409586905300167, |
| "learning_rate": 7.67102970754101e-06, |
| "loss": 0.7771, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.8071825566481402, |
| "grad_norm": 1.3689064689741741, |
| "learning_rate": 7.568906530820281e-06, |
| "loss": 0.7964, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.8208636169303121, |
| "grad_norm": 1.4659172981797304, |
| "learning_rate": 7.465305999860728e-06, |
| "loss": 0.7944, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.8345446772124839, |
| "grad_norm": 1.5394174805450438, |
| "learning_rate": 7.360287694363566e-06, |
| "loss": 0.8467, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.8482257374946558, |
| "grad_norm": 1.577909399946238, |
| "learning_rate": 7.253912009378953e-06, |
| "loss": 0.8211, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.8619067977768277, |
| "grad_norm": 1.1486399152909672, |
| "learning_rate": 7.146240120573358e-06, |
| "loss": 0.6165, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.8755878580589995, |
| "grad_norm": 1.3636010924515631, |
| "learning_rate": 7.037333949048005e-06, |
| "loss": 0.7381, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.8892689183411714, |
| "grad_norm": 1.275910605216143, |
| "learning_rate": 6.927256125728624e-06, |
| "loss": 0.7358, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.9029499786233434, |
| "grad_norm": 1.29369042675824, |
| "learning_rate": 6.816069955346986e-06, |
| "loss": 0.6921, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.9166310389055152, |
| "grad_norm": 1.4240273029942732, |
| "learning_rate": 6.703839380034945e-06, |
| "loss": 0.779, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.9303120991876871, |
| "grad_norm": 1.3292854453457577, |
| "learning_rate": 6.590628942551909e-06, |
| "loss": 0.7228, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.943993159469859, |
| "grad_norm": 1.496692141287322, |
| "learning_rate": 6.476503749166903e-06, |
| "loss": 0.8125, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.9576742197520308, |
| "grad_norm": 1.5176216507128955, |
| "learning_rate": 6.36152943221656e-06, |
| "loss": 0.8534, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.9713552800342027, |
| "grad_norm": 1.302742485236075, |
| "learning_rate": 6.245772112360568e-06, |
| "loss": 0.6825, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.9850363403163745, |
| "grad_norm": 1.6413121562783546, |
| "learning_rate": 6.129298360556304e-06, |
| "loss": 0.9217, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.9987174005985464, |
| "grad_norm": 1.3894377677004774, |
| "learning_rate": 6.012175159774488e-06, |
| "loss": 0.8585, |
| "step": 73 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 146, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 73, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 9628983689216.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|