| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.026314635323012148, | |
| "eval_steps": 500, | |
| "global_step": 300, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 8.771545107670716e-05, | |
| "grad_norm": 28.291993022356824, | |
| "learning_rate": 4.385964912280702e-08, | |
| "loss": 0.9764, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.00017543090215341433, | |
| "grad_norm": 11.00431285069151, | |
| "learning_rate": 8.771929824561404e-08, | |
| "loss": 0.7373, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0002631463532301215, | |
| "grad_norm": 19.575902791602918, | |
| "learning_rate": 1.3157894736842107e-07, | |
| "loss": 0.92, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.00035086180430682866, | |
| "grad_norm": 28.862884630243123, | |
| "learning_rate": 1.7543859649122808e-07, | |
| "loss": 0.9196, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.0004385772553835358, | |
| "grad_norm": 15.982248327528751, | |
| "learning_rate": 2.192982456140351e-07, | |
| "loss": 0.8366, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.000526292706460243, | |
| "grad_norm": 31.85723876161732, | |
| "learning_rate": 2.6315789473684213e-07, | |
| "loss": 0.9335, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.0006140081575369502, | |
| "grad_norm": 21.310207454796295, | |
| "learning_rate": 3.070175438596491e-07, | |
| "loss": 0.8362, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.0007017236086136573, | |
| "grad_norm": 20.052830776823505, | |
| "learning_rate": 3.5087719298245616e-07, | |
| "loss": 0.8015, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.0007894390596903645, | |
| "grad_norm": 16.06788143210757, | |
| "learning_rate": 3.9473684210526315e-07, | |
| "loss": 0.8729, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.0008771545107670716, | |
| "grad_norm": 29.100726513914584, | |
| "learning_rate": 4.385964912280702e-07, | |
| "loss": 0.9058, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0009648699618437788, | |
| "grad_norm": 13.993390572028792, | |
| "learning_rate": 4.824561403508772e-07, | |
| "loss": 0.7093, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.001052585412920486, | |
| "grad_norm": 21.107935511000072, | |
| "learning_rate": 5.263157894736843e-07, | |
| "loss": 0.8955, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.0011403008639971931, | |
| "grad_norm": 13.66193898339087, | |
| "learning_rate": 5.701754385964912e-07, | |
| "loss": 0.7219, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.0012280163150739003, | |
| "grad_norm": 10.537203866107753, | |
| "learning_rate": 6.140350877192982e-07, | |
| "loss": 0.8429, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.0013157317661506075, | |
| "grad_norm": 12.393106853157317, | |
| "learning_rate": 6.578947368421053e-07, | |
| "loss": 0.6708, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.0014034472172273146, | |
| "grad_norm": 8.734604355126535, | |
| "learning_rate": 7.017543859649123e-07, | |
| "loss": 0.6507, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.0014911626683040218, | |
| "grad_norm": 9.124362491394539, | |
| "learning_rate": 7.456140350877194e-07, | |
| "loss": 0.838, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.001578878119380729, | |
| "grad_norm": 8.958389642999963, | |
| "learning_rate": 7.894736842105263e-07, | |
| "loss": 0.6849, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.0016665935704574361, | |
| "grad_norm": 11.542677492312867, | |
| "learning_rate": 8.333333333333333e-07, | |
| "loss": 0.6926, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.0017543090215341433, | |
| "grad_norm": 8.045066225626593, | |
| "learning_rate": 8.771929824561404e-07, | |
| "loss": 0.7006, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0018420244726108505, | |
| "grad_norm": 8.146906074379428, | |
| "learning_rate": 9.210526315789474e-07, | |
| "loss": 0.6737, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.0019297399236875576, | |
| "grad_norm": 6.502955757535831, | |
| "learning_rate": 9.649122807017545e-07, | |
| "loss": 0.7495, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.002017455374764265, | |
| "grad_norm": 8.736982858234592, | |
| "learning_rate": 1.0087719298245615e-06, | |
| "loss": 0.7324, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.002105170825840972, | |
| "grad_norm": 7.851959741269017, | |
| "learning_rate": 1.0526315789473685e-06, | |
| "loss": 0.6686, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.002192886276917679, | |
| "grad_norm": 8.594840793358543, | |
| "learning_rate": 1.0964912280701756e-06, | |
| "loss": 0.8064, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.0022806017279943863, | |
| "grad_norm": 8.935665287337994, | |
| "learning_rate": 1.1403508771929824e-06, | |
| "loss": 0.6751, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.0023683171790710934, | |
| "grad_norm": 11.146850280588064, | |
| "learning_rate": 1.1842105263157894e-06, | |
| "loss": 0.7884, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.0024560326301478006, | |
| "grad_norm": 6.917869007862471, | |
| "learning_rate": 1.2280701754385965e-06, | |
| "loss": 0.8772, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.0025437480812245078, | |
| "grad_norm": 9.32145567192897, | |
| "learning_rate": 1.2719298245614037e-06, | |
| "loss": 0.6486, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.002631463532301215, | |
| "grad_norm": 7.83399807213587, | |
| "learning_rate": 1.3157894736842106e-06, | |
| "loss": 0.7793, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.002719178983377922, | |
| "grad_norm": 5.701851482721999, | |
| "learning_rate": 1.3596491228070178e-06, | |
| "loss": 0.6418, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.0028068944344546293, | |
| "grad_norm": 6.357569510522249, | |
| "learning_rate": 1.4035087719298246e-06, | |
| "loss": 0.7803, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.0028946098855313364, | |
| "grad_norm": 6.1458878660724, | |
| "learning_rate": 1.4473684210526317e-06, | |
| "loss": 0.6075, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.0029823253366080436, | |
| "grad_norm": 5.258525934759675, | |
| "learning_rate": 1.4912280701754387e-06, | |
| "loss": 0.7558, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.0030700407876847508, | |
| "grad_norm": 5.96497463401995, | |
| "learning_rate": 1.5350877192982458e-06, | |
| "loss": 0.5807, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.003157756238761458, | |
| "grad_norm": 9.97378904781871, | |
| "learning_rate": 1.5789473684210526e-06, | |
| "loss": 0.6766, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.003245471689838165, | |
| "grad_norm": 10.558130153122322, | |
| "learning_rate": 1.6228070175438598e-06, | |
| "loss": 0.6318, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.0033331871409148723, | |
| "grad_norm": 7.730592682668347, | |
| "learning_rate": 1.6666666666666667e-06, | |
| "loss": 0.5723, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.0034209025919915794, | |
| "grad_norm": 6.513997535111305, | |
| "learning_rate": 1.710526315789474e-06, | |
| "loss": 0.7381, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.0035086180430682866, | |
| "grad_norm": 6.4186997859745185, | |
| "learning_rate": 1.7543859649122807e-06, | |
| "loss": 0.676, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0035963334941449938, | |
| "grad_norm": 4.789756704738587, | |
| "learning_rate": 1.798245614035088e-06, | |
| "loss": 0.8106, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.003684048945221701, | |
| "grad_norm": 8.552415866186008, | |
| "learning_rate": 1.8421052631578948e-06, | |
| "loss": 0.7834, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.003771764396298408, | |
| "grad_norm": 5.104236885105078, | |
| "learning_rate": 1.8859649122807019e-06, | |
| "loss": 0.6694, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.0038594798473751152, | |
| "grad_norm": 6.998642641947579, | |
| "learning_rate": 1.929824561403509e-06, | |
| "loss": 0.7184, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.003947195298451822, | |
| "grad_norm": 6.754484565741454, | |
| "learning_rate": 1.973684210526316e-06, | |
| "loss": 0.7682, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.00403491074952853, | |
| "grad_norm": 5.702466747706841, | |
| "learning_rate": 2.017543859649123e-06, | |
| "loss": 0.7167, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.004122626200605236, | |
| "grad_norm": 7.038100758557257, | |
| "learning_rate": 2.06140350877193e-06, | |
| "loss": 0.6709, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.004210341651681944, | |
| "grad_norm": 8.659378609826204, | |
| "learning_rate": 2.105263157894737e-06, | |
| "loss": 0.6508, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.004298057102758651, | |
| "grad_norm": 9.315174303463822, | |
| "learning_rate": 2.149122807017544e-06, | |
| "loss": 0.6168, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.004385772553835358, | |
| "grad_norm": 7.447716885721135, | |
| "learning_rate": 2.192982456140351e-06, | |
| "loss": 0.6738, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.004473488004912065, | |
| "grad_norm": 5.600770404460154, | |
| "learning_rate": 2.236842105263158e-06, | |
| "loss": 0.6311, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.004561203455988773, | |
| "grad_norm": 7.059691201242354, | |
| "learning_rate": 2.280701754385965e-06, | |
| "loss": 0.7204, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.004648918907065479, | |
| "grad_norm": 5.589092290239263, | |
| "learning_rate": 2.324561403508772e-06, | |
| "loss": 0.7266, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.004736634358142187, | |
| "grad_norm": 5.801762781587569, | |
| "learning_rate": 2.368421052631579e-06, | |
| "loss": 0.5336, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.004824349809218894, | |
| "grad_norm": 5.599754768073974, | |
| "learning_rate": 2.412280701754386e-06, | |
| "loss": 0.6338, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.004912065260295601, | |
| "grad_norm": 5.66437398031977, | |
| "learning_rate": 2.456140350877193e-06, | |
| "loss": 0.7813, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.004999780711372308, | |
| "grad_norm": 6.32022790188225, | |
| "learning_rate": 2.5e-06, | |
| "loss": 0.6613, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.0050874961624490156, | |
| "grad_norm": 8.01474270706056, | |
| "learning_rate": 2.5438596491228075e-06, | |
| "loss": 0.6451, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.005175211613525722, | |
| "grad_norm": 6.586182462850705, | |
| "learning_rate": 2.5877192982456147e-06, | |
| "loss": 0.6984, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.00526292706460243, | |
| "grad_norm": 5.61553252576188, | |
| "learning_rate": 2.631578947368421e-06, | |
| "loss": 0.5773, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.005350642515679137, | |
| "grad_norm": 5.5274818204706895, | |
| "learning_rate": 2.6754385964912284e-06, | |
| "loss": 0.6083, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.005438357966755844, | |
| "grad_norm": 3.8762804528384254, | |
| "learning_rate": 2.7192982456140356e-06, | |
| "loss": 0.7174, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.005526073417832551, | |
| "grad_norm": 5.248404081335598, | |
| "learning_rate": 2.7631578947368424e-06, | |
| "loss": 0.7066, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.0056137888689092585, | |
| "grad_norm": 7.214109517049078, | |
| "learning_rate": 2.8070175438596493e-06, | |
| "loss": 0.692, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.005701504319985965, | |
| "grad_norm": 5.429278596290352, | |
| "learning_rate": 2.8508771929824565e-06, | |
| "loss": 0.6145, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.005789219771062673, | |
| "grad_norm": 17.638205100824422, | |
| "learning_rate": 2.8947368421052634e-06, | |
| "loss": 0.7677, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.00587693522213938, | |
| "grad_norm": 5.677374136021176, | |
| "learning_rate": 2.9385964912280706e-06, | |
| "loss": 0.6779, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.005964650673216087, | |
| "grad_norm": 5.453107411280262, | |
| "learning_rate": 2.9824561403508774e-06, | |
| "loss": 0.6428, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.006052366124292794, | |
| "grad_norm": 5.888626008478417, | |
| "learning_rate": 3.0263157894736843e-06, | |
| "loss": 0.6342, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.0061400815753695015, | |
| "grad_norm": 5.3185045733144225, | |
| "learning_rate": 3.0701754385964915e-06, | |
| "loss": 0.5644, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.006227797026446208, | |
| "grad_norm": 4.902919731780363, | |
| "learning_rate": 3.1140350877192988e-06, | |
| "loss": 0.709, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.006315512477522916, | |
| "grad_norm": 8.773622618503456, | |
| "learning_rate": 3.157894736842105e-06, | |
| "loss": 0.6674, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.006403227928599623, | |
| "grad_norm": 6.7570883776978174, | |
| "learning_rate": 3.2017543859649124e-06, | |
| "loss": 0.6918, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.00649094337967633, | |
| "grad_norm": 5.597179964370573, | |
| "learning_rate": 3.2456140350877197e-06, | |
| "loss": 0.7119, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.006578658830753037, | |
| "grad_norm": 5.4824260737552795, | |
| "learning_rate": 3.289473684210527e-06, | |
| "loss": 0.5667, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.0066663742818297445, | |
| "grad_norm": 6.083422094529157, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 0.5972, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.006754089732906451, | |
| "grad_norm": 6.688559230122185, | |
| "learning_rate": 3.3771929824561406e-06, | |
| "loss": 0.6079, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.006841805183983159, | |
| "grad_norm": 4.675152512564395, | |
| "learning_rate": 3.421052631578948e-06, | |
| "loss": 0.6431, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.006929520635059866, | |
| "grad_norm": 6.61824094926871, | |
| "learning_rate": 3.464912280701755e-06, | |
| "loss": 0.7219, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.007017236086136573, | |
| "grad_norm": 4.3090639659166685, | |
| "learning_rate": 3.5087719298245615e-06, | |
| "loss": 0.6267, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.00710495153721328, | |
| "grad_norm": 5.908526205124108, | |
| "learning_rate": 3.5526315789473687e-06, | |
| "loss": 0.5598, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.0071926669882899875, | |
| "grad_norm": 4.954945711406169, | |
| "learning_rate": 3.596491228070176e-06, | |
| "loss": 0.6251, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.007280382439366694, | |
| "grad_norm": 6.403352381905709, | |
| "learning_rate": 3.640350877192983e-06, | |
| "loss": 0.6921, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.007368097890443402, | |
| "grad_norm": 5.8960340556018505, | |
| "learning_rate": 3.6842105263157896e-06, | |
| "loss": 0.5803, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.007455813341520109, | |
| "grad_norm": 5.5832723717085795, | |
| "learning_rate": 3.728070175438597e-06, | |
| "loss": 0.7109, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.007543528792596816, | |
| "grad_norm": 6.9538610646678425, | |
| "learning_rate": 3.7719298245614037e-06, | |
| "loss": 0.57, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.007631244243673523, | |
| "grad_norm": 4.9040721673618615, | |
| "learning_rate": 3.815789473684211e-06, | |
| "loss": 0.6681, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.0077189596947502305, | |
| "grad_norm": 4.367227562952691, | |
| "learning_rate": 3.859649122807018e-06, | |
| "loss": 0.5881, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.007806675145826937, | |
| "grad_norm": 6.135869823936115, | |
| "learning_rate": 3.903508771929825e-06, | |
| "loss": 0.6333, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.007894390596903644, | |
| "grad_norm": 5.26232269598073, | |
| "learning_rate": 3.947368421052632e-06, | |
| "loss": 0.6228, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.007982106047980352, | |
| "grad_norm": 5.478510766614749, | |
| "learning_rate": 3.991228070175439e-06, | |
| "loss": 0.6889, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.00806982149905706, | |
| "grad_norm": 7.252221492478827, | |
| "learning_rate": 4.035087719298246e-06, | |
| "loss": 0.6726, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.008157536950133767, | |
| "grad_norm": 6.810323867433885, | |
| "learning_rate": 4.078947368421053e-06, | |
| "loss": 0.6186, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.008245252401210473, | |
| "grad_norm": 5.1477310672971965, | |
| "learning_rate": 4.12280701754386e-06, | |
| "loss": 0.6739, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.00833296785228718, | |
| "grad_norm": 4.455009313283226, | |
| "learning_rate": 4.166666666666667e-06, | |
| "loss": 0.6676, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.008420683303363888, | |
| "grad_norm": 4.854476484535793, | |
| "learning_rate": 4.210526315789474e-06, | |
| "loss": 0.624, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.008508398754440595, | |
| "grad_norm": 8.775528791539337, | |
| "learning_rate": 4.254385964912281e-06, | |
| "loss": 0.7236, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.008596114205517301, | |
| "grad_norm": 4.656928105654083, | |
| "learning_rate": 4.298245614035088e-06, | |
| "loss": 0.4853, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.008683829656594009, | |
| "grad_norm": 6.1151229878888795, | |
| "learning_rate": 4.342105263157895e-06, | |
| "loss": 0.6611, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.008771545107670716, | |
| "grad_norm": 4.846266795088099, | |
| "learning_rate": 4.385964912280702e-06, | |
| "loss": 0.6899, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.008859260558747424, | |
| "grad_norm": 5.63076019856985, | |
| "learning_rate": 4.429824561403509e-06, | |
| "loss": 0.7394, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.00894697600982413, | |
| "grad_norm": 6.152211661702361, | |
| "learning_rate": 4.473684210526316e-06, | |
| "loss": 0.6366, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.009034691460900838, | |
| "grad_norm": 5.271237730819475, | |
| "learning_rate": 4.517543859649123e-06, | |
| "loss": 0.6776, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.009122406911977545, | |
| "grad_norm": 6.150704296921181, | |
| "learning_rate": 4.56140350877193e-06, | |
| "loss": 0.7287, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.009210122363054253, | |
| "grad_norm": 5.511353295743786, | |
| "learning_rate": 4.605263157894737e-06, | |
| "loss": 0.7156, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.009297837814130959, | |
| "grad_norm": 5.651321362023493, | |
| "learning_rate": 4.649122807017544e-06, | |
| "loss": 0.5971, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.009385553265207666, | |
| "grad_norm": 4.521052312786367, | |
| "learning_rate": 4.692982456140351e-06, | |
| "loss": 0.662, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.009473268716284374, | |
| "grad_norm": 6.5893774516601775, | |
| "learning_rate": 4.736842105263158e-06, | |
| "loss": 0.6838, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.009560984167361081, | |
| "grad_norm": 7.413604525506308, | |
| "learning_rate": 4.780701754385965e-06, | |
| "loss": 0.6798, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.009648699618437787, | |
| "grad_norm": 5.258683042524991, | |
| "learning_rate": 4.824561403508772e-06, | |
| "loss": 0.7137, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.009736415069514495, | |
| "grad_norm": 3.56629655229689, | |
| "learning_rate": 4.8684210526315795e-06, | |
| "loss": 0.5524, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.009824130520591202, | |
| "grad_norm": 7.972594797604, | |
| "learning_rate": 4.912280701754386e-06, | |
| "loss": 0.7946, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.00991184597166791, | |
| "grad_norm": 5.9169587346561965, | |
| "learning_rate": 4.956140350877193e-06, | |
| "loss": 0.6985, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.009999561422744616, | |
| "grad_norm": 4.9028768240583895, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7471, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.010087276873821324, | |
| "grad_norm": 4.952040118758915, | |
| "learning_rate": 4.999999903143301e-06, | |
| "loss": 0.6645, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.010174992324898031, | |
| "grad_norm": 5.307375041926707, | |
| "learning_rate": 4.999999612573212e-06, | |
| "loss": 0.6568, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.010262707775974739, | |
| "grad_norm": 4.417210142946582, | |
| "learning_rate": 4.9999991282897545e-06, | |
| "loss": 0.6633, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.010350423227051445, | |
| "grad_norm": 6.813103500844099, | |
| "learning_rate": 4.999998450292966e-06, | |
| "loss": 0.7479, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.010438138678128152, | |
| "grad_norm": 5.220452049535287, | |
| "learning_rate": 4.9999975785829e-06, | |
| "loss": 0.5982, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.01052585412920486, | |
| "grad_norm": 6.470241976711781, | |
| "learning_rate": 4.999996513159624e-06, | |
| "loss": 0.5915, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.010613569580281567, | |
| "grad_norm": 5.236784827517624, | |
| "learning_rate": 4.99999525402322e-06, | |
| "loss": 0.665, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.010701285031358273, | |
| "grad_norm": 5.5322906674158565, | |
| "learning_rate": 4.999993801173785e-06, | |
| "loss": 0.473, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.01078900048243498, | |
| "grad_norm": 5.643434680672429, | |
| "learning_rate": 4.999992154611433e-06, | |
| "loss": 0.5802, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.010876715933511688, | |
| "grad_norm": 4.909123022379139, | |
| "learning_rate": 4.9999903143362905e-06, | |
| "loss": 0.6103, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.010964431384588396, | |
| "grad_norm": 7.046173121098522, | |
| "learning_rate": 4.999988280348501e-06, | |
| "loss": 0.6601, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.011052146835665102, | |
| "grad_norm": 5.567754476589664, | |
| "learning_rate": 4.99998605264822e-06, | |
| "loss": 0.7144, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.01113986228674181, | |
| "grad_norm": 6.670512866037107, | |
| "learning_rate": 4.999983631235623e-06, | |
| "loss": 0.5034, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.011227577737818517, | |
| "grad_norm": 5.068760146843144, | |
| "learning_rate": 4.999981016110896e-06, | |
| "loss": 0.5965, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.011315293188895225, | |
| "grad_norm": 5.493410339028754, | |
| "learning_rate": 4.999978207274243e-06, | |
| "loss": 0.6697, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.01140300863997193, | |
| "grad_norm": 5.662089015796081, | |
| "learning_rate": 4.999975204725879e-06, | |
| "loss": 0.7182, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.011490724091048638, | |
| "grad_norm": 3.734356064938746, | |
| "learning_rate": 4.999972008466039e-06, | |
| "loss": 0.632, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.011578439542125346, | |
| "grad_norm": 4.29907687663725, | |
| "learning_rate": 4.99996861849497e-06, | |
| "loss": 0.6321, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.011666154993202053, | |
| "grad_norm": 5.292963722155827, | |
| "learning_rate": 4.999965034812934e-06, | |
| "loss": 0.5768, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.01175387044427876, | |
| "grad_norm": 4.564589196086129, | |
| "learning_rate": 4.99996125742021e-06, | |
| "loss": 0.5991, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.011841585895355467, | |
| "grad_norm": 5.889974426321806, | |
| "learning_rate": 4.99995728631709e-06, | |
| "loss": 0.568, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.011929301346432174, | |
| "grad_norm": 4.903556688362067, | |
| "learning_rate": 4.999953121503881e-06, | |
| "loss": 0.6221, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.012017016797508882, | |
| "grad_norm": 4.652137494582458, | |
| "learning_rate": 4.999948762980906e-06, | |
| "loss": 0.6499, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.012104732248585588, | |
| "grad_norm": 7.2681565015460965, | |
| "learning_rate": 4.999944210748504e-06, | |
| "loss": 0.7997, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.012192447699662295, | |
| "grad_norm": 4.498966830496647, | |
| "learning_rate": 4.999939464807027e-06, | |
| "loss": 0.7033, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.012280163150739003, | |
| "grad_norm": 5.658829625864849, | |
| "learning_rate": 4.999934525156842e-06, | |
| "loss": 0.6234, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.01236787860181571, | |
| "grad_norm": 6.170987539440289, | |
| "learning_rate": 4.9999293917983325e-06, | |
| "loss": 0.7359, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.012455594052892417, | |
| "grad_norm": 4.889450035742974, | |
| "learning_rate": 4.999924064731896e-06, | |
| "loss": 0.6418, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.012543309503969124, | |
| "grad_norm": 5.565665252735285, | |
| "learning_rate": 4.9999185439579445e-06, | |
| "loss": 0.8114, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.012631024955045832, | |
| "grad_norm": 5.009655972578068, | |
| "learning_rate": 4.9999128294769075e-06, | |
| "loss": 0.7307, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.01271874040612254, | |
| "grad_norm": 5.011444448419762, | |
| "learning_rate": 4.999906921289227e-06, | |
| "loss": 0.6434, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.012806455857199245, | |
| "grad_norm": 5.91290249112379, | |
| "learning_rate": 4.999900819395361e-06, | |
| "loss": 0.7576, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.012894171308275953, | |
| "grad_norm": 5.291827066915767, | |
| "learning_rate": 4.9998945237957814e-06, | |
| "loss": 0.717, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.01298188675935266, | |
| "grad_norm": 6.889695918810895, | |
| "learning_rate": 4.9998880344909765e-06, | |
| "loss": 0.6566, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.013069602210429368, | |
| "grad_norm": 4.139725258131711, | |
| "learning_rate": 4.999881351481449e-06, | |
| "loss": 0.6139, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.013157317661506074, | |
| "grad_norm": 5.041147601092224, | |
| "learning_rate": 4.999874474767718e-06, | |
| "loss": 0.7046, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.013245033112582781, | |
| "grad_norm": 4.850191233243735, | |
| "learning_rate": 4.999867404350315e-06, | |
| "loss": 0.6494, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.013332748563659489, | |
| "grad_norm": 5.608814210289025, | |
| "learning_rate": 4.999860140229788e-06, | |
| "loss": 0.8654, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.013420464014736197, | |
| "grad_norm": 4.097824317856954, | |
| "learning_rate": 4.9998526824067e-06, | |
| "loss": 0.6889, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.013508179465812903, | |
| "grad_norm": 6.425927321695068, | |
| "learning_rate": 4.999845030881629e-06, | |
| "loss": 0.5837, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.01359589491688961, | |
| "grad_norm": 7.686652681051417, | |
| "learning_rate": 4.999837185655168e-06, | |
| "loss": 0.6869, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.013683610367966318, | |
| "grad_norm": 6.199666417167642, | |
| "learning_rate": 4.9998291467279245e-06, | |
| "loss": 0.7371, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.013771325819043024, | |
| "grad_norm": 6.797879751043678, | |
| "learning_rate": 4.999820914100522e-06, | |
| "loss": 0.6912, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.013859041270119731, | |
| "grad_norm": 9.837640179642968, | |
| "learning_rate": 4.999812487773597e-06, | |
| "loss": 0.8045, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.013946756721196439, | |
| "grad_norm": 6.620454193744729, | |
| "learning_rate": 4.9998038677478044e-06, | |
| "loss": 0.6018, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.014034472172273146, | |
| "grad_norm": 4.952380418390811, | |
| "learning_rate": 4.99979505402381e-06, | |
| "loss": 0.5851, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.014122187623349852, | |
| "grad_norm": 4.571346505498035, | |
| "learning_rate": 4.999786046602299e-06, | |
| "loss": 0.6633, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.01420990307442656, | |
| "grad_norm": 6.745466717777739, | |
| "learning_rate": 4.999776845483968e-06, | |
| "loss": 0.714, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.014297618525503267, | |
| "grad_norm": 4.888639355192875, | |
| "learning_rate": 4.999767450669531e-06, | |
| "loss": 0.5328, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.014385333976579975, | |
| "grad_norm": 5.263414218540685, | |
| "learning_rate": 4.999757862159713e-06, | |
| "loss": 0.6746, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.014473049427656681, | |
| "grad_norm": 5.8723140369149895, | |
| "learning_rate": 4.99974807995526e-06, | |
| "loss": 0.7101, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.014560764878733388, | |
| "grad_norm": 4.125348885535371, | |
| "learning_rate": 4.999738104056931e-06, | |
| "loss": 0.6418, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.014648480329810096, | |
| "grad_norm": 5.079939786355144, | |
| "learning_rate": 4.999727934465495e-06, | |
| "loss": 0.6757, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.014736195780886804, | |
| "grad_norm": 4.436648943550616, | |
| "learning_rate": 4.999717571181742e-06, | |
| "loss": 0.6878, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.01482391123196351, | |
| "grad_norm": 4.6070293178483706, | |
| "learning_rate": 4.999707014206475e-06, | |
| "loss": 0.6882, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.014911626683040217, | |
| "grad_norm": 4.337658765605819, | |
| "learning_rate": 4.999696263540513e-06, | |
| "loss": 0.6418, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.014999342134116925, | |
| "grad_norm": 5.834498841218243, | |
| "learning_rate": 4.999685319184688e-06, | |
| "loss": 0.6367, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.015087057585193632, | |
| "grad_norm": 6.027148776110112, | |
| "learning_rate": 4.999674181139848e-06, | |
| "loss": 0.7505, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.015174773036270338, | |
| "grad_norm": 4.712652033599274, | |
| "learning_rate": 4.999662849406855e-06, | |
| "loss": 0.7515, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.015262488487347046, | |
| "grad_norm": 5.325275991673836, | |
| "learning_rate": 4.99965132398659e-06, | |
| "loss": 0.7871, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.015350203938423753, | |
| "grad_norm": 5.006048437293231, | |
| "learning_rate": 4.999639604879943e-06, | |
| "loss": 0.6038, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.015437919389500461, | |
| "grad_norm": 4.692976251794895, | |
| "learning_rate": 4.999627692087824e-06, | |
| "loss": 0.7106, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.015525634840577167, | |
| "grad_norm": 6.484912012474024, | |
| "learning_rate": 4.999615585611156e-06, | |
| "loss": 0.6456, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.015613350291653874, | |
| "grad_norm": 7.072312221146792, | |
| "learning_rate": 4.999603285450875e-06, | |
| "loss": 0.6986, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.015701065742730582, | |
| "grad_norm": 5.072158684292459, | |
| "learning_rate": 4.999590791607936e-06, | |
| "loss": 0.6386, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.015788781193807288, | |
| "grad_norm": 5.674801641765509, | |
| "learning_rate": 4.999578104083307e-06, | |
| "loss": 0.6512, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.015876496644883997, | |
| "grad_norm": 6.011232915930249, | |
| "learning_rate": 4.9995652228779715e-06, | |
| "loss": 0.6166, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.015964212095960703, | |
| "grad_norm": 7.067996556252431, | |
| "learning_rate": 4.999552147992926e-06, | |
| "loss": 0.8316, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.01605192754703741, | |
| "grad_norm": 6.191586224655665, | |
| "learning_rate": 4.999538879429183e-06, | |
| "loss": 0.7167, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.01613964299811412, | |
| "grad_norm": 5.40861794404673, | |
| "learning_rate": 4.999525417187774e-06, | |
| "loss": 0.6604, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.016227358449190824, | |
| "grad_norm": 5.619694849325643, | |
| "learning_rate": 4.999511761269739e-06, | |
| "loss": 0.7141, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.016315073900267534, | |
| "grad_norm": 7.467663008400906, | |
| "learning_rate": 4.999497911676138e-06, | |
| "loss": 0.6086, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.01640278935134424, | |
| "grad_norm": 4.645589903763359, | |
| "learning_rate": 4.999483868408043e-06, | |
| "loss": 0.6932, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.016490504802420945, | |
| "grad_norm": 4.819294533224638, | |
| "learning_rate": 4.999469631466544e-06, | |
| "loss": 0.6256, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.016578220253497655, | |
| "grad_norm": 4.711171445741636, | |
| "learning_rate": 4.999455200852741e-06, | |
| "loss": 0.7445, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.01666593570457436, | |
| "grad_norm": 4.371758877075776, | |
| "learning_rate": 4.999440576567755e-06, | |
| "loss": 0.6801, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.016753651155651066, | |
| "grad_norm": 5.761171404408883, | |
| "learning_rate": 4.999425758612718e-06, | |
| "loss": 0.6701, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.016841366606727776, | |
| "grad_norm": 4.340375314807721, | |
| "learning_rate": 4.999410746988778e-06, | |
| "loss": 0.5556, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.01692908205780448, | |
| "grad_norm": 4.775058922031801, | |
| "learning_rate": 4.9993955416970986e-06, | |
| "loss": 0.6915, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.01701679750888119, | |
| "grad_norm": 4.301940379009061, | |
| "learning_rate": 4.999380142738857e-06, | |
| "loss": 0.6982, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.017104512959957897, | |
| "grad_norm": 4.746670538298819, | |
| "learning_rate": 4.9993645501152485e-06, | |
| "loss": 0.5392, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.017192228411034603, | |
| "grad_norm": 5.312812102176541, | |
| "learning_rate": 4.999348763827479e-06, | |
| "loss": 0.6254, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.017279943862111312, | |
| "grad_norm": 6.073252701324542, | |
| "learning_rate": 4.999332783876774e-06, | |
| "loss": 0.7221, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.017367659313188018, | |
| "grad_norm": 6.783014797465277, | |
| "learning_rate": 4.999316610264369e-06, | |
| "loss": 0.5914, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.017455374764264724, | |
| "grad_norm": 5.105373260000072, | |
| "learning_rate": 4.999300242991519e-06, | |
| "loss": 0.4895, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.017543090215341433, | |
| "grad_norm": 5.3256898167081825, | |
| "learning_rate": 4.999283682059493e-06, | |
| "loss": 0.714, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.01763080566641814, | |
| "grad_norm": 7.815945435660424, | |
| "learning_rate": 4.999266927469572e-06, | |
| "loss": 0.7691, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.017718521117494848, | |
| "grad_norm": 4.350216346007481, | |
| "learning_rate": 4.999249979223056e-06, | |
| "loss": 0.7205, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.017806236568571554, | |
| "grad_norm": 4.167534183562087, | |
| "learning_rate": 4.999232837321257e-06, | |
| "loss": 0.6716, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.01789395201964826, | |
| "grad_norm": 6.564156035042191, | |
| "learning_rate": 4.999215501765504e-06, | |
| "loss": 0.6139, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.01798166747072497, | |
| "grad_norm": 4.58988335300785, | |
| "learning_rate": 4.9991979725571395e-06, | |
| "loss": 0.6241, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.018069382921801675, | |
| "grad_norm": 7.14774553510386, | |
| "learning_rate": 4.999180249697524e-06, | |
| "loss": 0.7338, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.01815709837287838, | |
| "grad_norm": 4.3154768710391656, | |
| "learning_rate": 4.999162333188028e-06, | |
| "loss": 0.646, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.01824481382395509, | |
| "grad_norm": 3.930924147546703, | |
| "learning_rate": 4.999144223030041e-06, | |
| "loss": 0.7162, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.018332529275031796, | |
| "grad_norm": 3.75066761929553, | |
| "learning_rate": 4.999125919224966e-06, | |
| "loss": 0.6283, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.018420244726108505, | |
| "grad_norm": 4.916459254987505, | |
| "learning_rate": 4.999107421774222e-06, | |
| "loss": 0.6716, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.01850796017718521, | |
| "grad_norm": 4.570226928027306, | |
| "learning_rate": 4.999088730679241e-06, | |
| "loss": 0.6527, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.018595675628261917, | |
| "grad_norm": 3.6658012035372605, | |
| "learning_rate": 4.999069845941472e-06, | |
| "loss": 0.5452, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.018683391079338627, | |
| "grad_norm": 4.697816375671605, | |
| "learning_rate": 4.999050767562379e-06, | |
| "loss": 0.7316, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.018771106530415332, | |
| "grad_norm": 5.639876519194002, | |
| "learning_rate": 4.99903149554344e-06, | |
| "loss": 0.5152, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.018858821981492038, | |
| "grad_norm": 5.527702869650481, | |
| "learning_rate": 4.999012029886147e-06, | |
| "loss": 0.6119, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.018946537432568748, | |
| "grad_norm": 6.019639388484205, | |
| "learning_rate": 4.998992370592008e-06, | |
| "loss": 0.7366, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.019034252883645453, | |
| "grad_norm": 4.014799337285965, | |
| "learning_rate": 4.998972517662549e-06, | |
| "loss": 0.7088, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.019121968334722163, | |
| "grad_norm": 7.876499612097003, | |
| "learning_rate": 4.998952471099307e-06, | |
| "loss": 0.5565, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.01920968378579887, | |
| "grad_norm": 7.386792956892447, | |
| "learning_rate": 4.998932230903835e-06, | |
| "loss": 0.6387, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.019297399236875575, | |
| "grad_norm": 5.346097163630257, | |
| "learning_rate": 4.998911797077701e-06, | |
| "loss": 0.6237, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.019385114687952284, | |
| "grad_norm": 6.133310652425816, | |
| "learning_rate": 4.998891169622488e-06, | |
| "loss": 0.7428, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.01947283013902899, | |
| "grad_norm": 4.224801633855712, | |
| "learning_rate": 4.998870348539797e-06, | |
| "loss": 0.7206, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.019560545590105696, | |
| "grad_norm": 5.648869005800134, | |
| "learning_rate": 4.998849333831238e-06, | |
| "loss": 0.6249, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.019648261041182405, | |
| "grad_norm": 4.634920959306503, | |
| "learning_rate": 4.998828125498441e-06, | |
| "loss": 0.6764, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.01973597649225911, | |
| "grad_norm": 4.882651557085375, | |
| "learning_rate": 4.998806723543049e-06, | |
| "loss": 0.6682, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.01982369194333582, | |
| "grad_norm": 4.5073631852916645, | |
| "learning_rate": 4.998785127966721e-06, | |
| "loss": 0.7658, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.019911407394412526, | |
| "grad_norm": 6.444404326993186, | |
| "learning_rate": 4.99876333877113e-06, | |
| "loss": 0.7161, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.019999122845489232, | |
| "grad_norm": 5.926254683053582, | |
| "learning_rate": 4.998741355957963e-06, | |
| "loss": 0.6083, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.02008683829656594, | |
| "grad_norm": 4.715935033600424, | |
| "learning_rate": 4.998719179528925e-06, | |
| "loss": 0.5764, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.020174553747642647, | |
| "grad_norm": 4.06642116262848, | |
| "learning_rate": 4.998696809485734e-06, | |
| "loss": 0.6436, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.020262269198719353, | |
| "grad_norm": 4.060536926809771, | |
| "learning_rate": 4.998674245830123e-06, | |
| "loss": 0.6455, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.020349984649796062, | |
| "grad_norm": 5.769596888340199, | |
| "learning_rate": 4.9986514885638405e-06, | |
| "loss": 0.6422, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.020437700100872768, | |
| "grad_norm": 5.619149975421577, | |
| "learning_rate": 4.99862853768865e-06, | |
| "loss": 0.5151, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.020525415551949477, | |
| "grad_norm": 5.738973149236573, | |
| "learning_rate": 4.998605393206329e-06, | |
| "loss": 0.5698, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.020613131003026183, | |
| "grad_norm": 3.9117936997485443, | |
| "learning_rate": 4.998582055118672e-06, | |
| "loss": 0.6139, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.02070084645410289, | |
| "grad_norm": 5.594946157519774, | |
| "learning_rate": 4.998558523427488e-06, | |
| "loss": 0.6305, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.0207885619051796, | |
| "grad_norm": 3.7796595114227816, | |
| "learning_rate": 4.998534798134598e-06, | |
| "loss": 0.6064, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.020876277356256304, | |
| "grad_norm": 5.530110712124758, | |
| "learning_rate": 4.998510879241842e-06, | |
| "loss": 0.7404, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.02096399280733301, | |
| "grad_norm": 5.795681054870311, | |
| "learning_rate": 4.998486766751073e-06, | |
| "loss": 0.6637, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.02105170825840972, | |
| "grad_norm": 5.250443330736557, | |
| "learning_rate": 4.99846246066416e-06, | |
| "loss": 0.7229, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.021139423709486425, | |
| "grad_norm": 5.307033877732376, | |
| "learning_rate": 4.998437960982985e-06, | |
| "loss": 0.729, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.021227139160563135, | |
| "grad_norm": 4.264326950314863, | |
| "learning_rate": 4.998413267709446e-06, | |
| "loss": 0.6363, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.02131485461163984, | |
| "grad_norm": 4.56674428695937, | |
| "learning_rate": 4.99838838084546e-06, | |
| "loss": 0.573, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.021402570062716546, | |
| "grad_norm": 5.367393577306364, | |
| "learning_rate": 4.998363300392951e-06, | |
| "loss": 0.6187, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.021490285513793256, | |
| "grad_norm": 5.58627031411974, | |
| "learning_rate": 4.998338026353865e-06, | |
| "loss": 0.635, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.02157800096486996, | |
| "grad_norm": 4.1536241104050005, | |
| "learning_rate": 4.9983125587301594e-06, | |
| "loss": 0.7296, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.021665716415946668, | |
| "grad_norm": 5.369955138376355, | |
| "learning_rate": 4.998286897523808e-06, | |
| "loss": 0.5939, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.021753431867023377, | |
| "grad_norm": 4.749169550030242, | |
| "learning_rate": 4.998261042736799e-06, | |
| "loss": 0.7125, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.021841147318100083, | |
| "grad_norm": 3.847851803716185, | |
| "learning_rate": 4.998234994371135e-06, | |
| "loss": 0.6874, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.021928862769176792, | |
| "grad_norm": 6.3610718821634755, | |
| "learning_rate": 4.998208752428836e-06, | |
| "loss": 0.6839, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.022016578220253498, | |
| "grad_norm": 6.90892255007994, | |
| "learning_rate": 4.998182316911934e-06, | |
| "loss": 0.6706, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.022104293671330204, | |
| "grad_norm": 4.842858396629252, | |
| "learning_rate": 4.998155687822478e-06, | |
| "loss": 0.7887, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.022192009122406913, | |
| "grad_norm": 6.80960196083629, | |
| "learning_rate": 4.99812886516253e-06, | |
| "loss": 0.6891, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.02227972457348362, | |
| "grad_norm": 6.897100992823047, | |
| "learning_rate": 4.998101848934171e-06, | |
| "loss": 0.7213, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.022367440024560325, | |
| "grad_norm": 4.383904436150581, | |
| "learning_rate": 4.9980746391394916e-06, | |
| "loss": 0.5472, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.022455155475637034, | |
| "grad_norm": 6.136102422729719, | |
| "learning_rate": 4.998047235780603e-06, | |
| "loss": 0.7462, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.02254287092671374, | |
| "grad_norm": 5.873462354540876, | |
| "learning_rate": 4.9980196388596255e-06, | |
| "loss": 0.6893, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.02263058637779045, | |
| "grad_norm": 5.36389164609212, | |
| "learning_rate": 4.9979918483787e-06, | |
| "loss": 0.725, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.022718301828867155, | |
| "grad_norm": 6.634852411669424, | |
| "learning_rate": 4.997963864339978e-06, | |
| "loss": 0.7619, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.02280601727994386, | |
| "grad_norm": 4.201015694891079, | |
| "learning_rate": 4.99793568674563e-06, | |
| "loss": 0.653, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.02289373273102057, | |
| "grad_norm": 4.951129353141893, | |
| "learning_rate": 4.997907315597836e-06, | |
| "loss": 0.7543, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.022981448182097276, | |
| "grad_norm": 4.331792323630216, | |
| "learning_rate": 4.997878750898798e-06, | |
| "loss": 0.6553, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.023069163633173982, | |
| "grad_norm": 4.764837636647203, | |
| "learning_rate": 4.997849992650727e-06, | |
| "loss": 0.719, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.02315687908425069, | |
| "grad_norm": 7.315146297212186, | |
| "learning_rate": 4.997821040855852e-06, | |
| "loss": 0.8217, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.023244594535327397, | |
| "grad_norm": 4.5164891139288015, | |
| "learning_rate": 4.997791895516417e-06, | |
| "loss": 0.5553, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.023332309986404107, | |
| "grad_norm": 4.651549875308793, | |
| "learning_rate": 4.99776255663468e-06, | |
| "loss": 0.6981, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.023420025437480813, | |
| "grad_norm": 4.941120481014187, | |
| "learning_rate": 4.997733024212913e-06, | |
| "loss": 0.604, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.02350774088855752, | |
| "grad_norm": 6.3616778757465315, | |
| "learning_rate": 4.997703298253406e-06, | |
| "loss": 0.7253, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.023595456339634228, | |
| "grad_norm": 4.723855693485358, | |
| "learning_rate": 4.997673378758462e-06, | |
| "loss": 0.7335, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.023683171790710934, | |
| "grad_norm": 4.336523073382538, | |
| "learning_rate": 4.997643265730399e-06, | |
| "loss": 0.5665, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.02377088724178764, | |
| "grad_norm": 6.547875149524498, | |
| "learning_rate": 4.997612959171549e-06, | |
| "loss": 0.6542, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.02385860269286435, | |
| "grad_norm": 5.285021138793967, | |
| "learning_rate": 4.997582459084264e-06, | |
| "loss": 0.7824, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.023946318143941055, | |
| "grad_norm": 4.447718203152539, | |
| "learning_rate": 4.9975517654709025e-06, | |
| "loss": 0.6728, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.024034033595017764, | |
| "grad_norm": 4.323105158596241, | |
| "learning_rate": 4.997520878333847e-06, | |
| "loss": 0.6516, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.02412174904609447, | |
| "grad_norm": 4.091596093860627, | |
| "learning_rate": 4.997489797675489e-06, | |
| "loss": 0.5786, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.024209464497171176, | |
| "grad_norm": 4.50262054947591, | |
| "learning_rate": 4.997458523498236e-06, | |
| "loss": 0.6632, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.024297179948247885, | |
| "grad_norm": 5.394966563241667, | |
| "learning_rate": 4.997427055804513e-06, | |
| "loss": 0.7415, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.02438489539932459, | |
| "grad_norm": 5.134838704391961, | |
| "learning_rate": 4.9973953945967565e-06, | |
| "loss": 0.6225, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.024472610850401297, | |
| "grad_norm": 4.555937935551801, | |
| "learning_rate": 4.9973635398774226e-06, | |
| "loss": 0.7451, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.024560326301478006, | |
| "grad_norm": 4.014041307501394, | |
| "learning_rate": 4.997331491648976e-06, | |
| "loss": 0.607, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.024648041752554712, | |
| "grad_norm": 5.398424400960683, | |
| "learning_rate": 4.9972992499139025e-06, | |
| "loss": 0.665, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.02473575720363142, | |
| "grad_norm": 6.959554022697295, | |
| "learning_rate": 4.9972668146746995e-06, | |
| "loss": 0.8175, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.024823472654708127, | |
| "grad_norm": 5.048396931572014, | |
| "learning_rate": 4.997234185933879e-06, | |
| "loss": 0.6961, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.024911188105784833, | |
| "grad_norm": 4.737474855724115, | |
| "learning_rate": 4.997201363693972e-06, | |
| "loss": 0.5337, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.024998903556861542, | |
| "grad_norm": 7.374843310231967, | |
| "learning_rate": 4.997168347957521e-06, | |
| "loss": 0.6791, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.025086619007938248, | |
| "grad_norm": 4.306967488515473, | |
| "learning_rate": 4.997135138727081e-06, | |
| "loss": 0.8791, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.025174334459014954, | |
| "grad_norm": 3.7949900410813737, | |
| "learning_rate": 4.99710173600523e-06, | |
| "loss": 0.7743, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.025262049910091663, | |
| "grad_norm": 4.842604758031469, | |
| "learning_rate": 4.997068139794554e-06, | |
| "loss": 0.6602, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.02534976536116837, | |
| "grad_norm": 3.531764677671023, | |
| "learning_rate": 4.9970343500976545e-06, | |
| "loss": 0.6317, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.02543748081224508, | |
| "grad_norm": 5.68234167540357, | |
| "learning_rate": 4.997000366917153e-06, | |
| "loss": 0.7404, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.025525196263321785, | |
| "grad_norm": 4.623883782994243, | |
| "learning_rate": 4.9969661902556804e-06, | |
| "loss": 0.6093, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.02561291171439849, | |
| "grad_norm": 5.9956405593570175, | |
| "learning_rate": 4.996931820115885e-06, | |
| "loss": 0.6773, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.0257006271654752, | |
| "grad_norm": 5.06274620174889, | |
| "learning_rate": 4.996897256500433e-06, | |
| "loss": 0.7249, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.025788342616551906, | |
| "grad_norm": 5.989915075597491, | |
| "learning_rate": 4.996862499411998e-06, | |
| "loss": 0.7526, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.02587605806762861, | |
| "grad_norm": 4.58567195302804, | |
| "learning_rate": 4.996827548853276e-06, | |
| "loss": 0.6762, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.02596377351870532, | |
| "grad_norm": 4.097368677404026, | |
| "learning_rate": 4.996792404826974e-06, | |
| "loss": 0.6238, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.026051488969782027, | |
| "grad_norm": 4.021749832913485, | |
| "learning_rate": 4.996757067335816e-06, | |
| "loss": 0.7958, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.026139204420858736, | |
| "grad_norm": 4.679522912267575, | |
| "learning_rate": 4.99672153638254e-06, | |
| "loss": 0.6583, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.026226919871935442, | |
| "grad_norm": 4.256974035317045, | |
| "learning_rate": 4.996685811969898e-06, | |
| "loss": 0.6464, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.026314635323012148, | |
| "grad_norm": 4.4862335847168096, | |
| "learning_rate": 4.996649894100659e-06, | |
| "loss": 0.6116, | |
| "step": 300 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 11400, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 150, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 7802380615680.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |