| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 49.99994058582378, | |
| "global_step": 420750, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 2.3624480095068333e-07, | |
| "loss": 2.5928, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.7391562685680333e-07, | |
| "loss": 1.753, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 7.115864527629234e-07, | |
| "loss": 1.5245, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 9.492572786690435e-07, | |
| "loss": 1.428, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 1.1869281045751634e-06, | |
| "loss": 1.3632, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.4245989304812835e-06, | |
| "loss": 1.3351, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.6622697563874036e-06, | |
| "loss": 1.3066, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.8999405822935234e-06, | |
| "loss": 1.266, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 2.137611408199644e-06, | |
| "loss": 1.2228, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.375282234105764e-06, | |
| "loss": 1.2047, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 2.6129530600118837e-06, | |
| "loss": 1.193, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 2.8501485442661915e-06, | |
| "loss": 1.169, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 3.0878193701723116e-06, | |
| "loss": 1.1512, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.325490196078431e-06, | |
| "loss": 1.1434, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.563161021984552e-06, | |
| "loss": 1.1305, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.8008318478906718e-06, | |
| "loss": 1.1059, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.038502673796792e-06, | |
| "loss": 1.1072, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 4.276173499702912e-06, | |
| "loss": 1.0856, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 4.5138443256090316e-06, | |
| "loss": 1.0605, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 4.751515151515152e-06, | |
| "loss": 1.0481, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 4.989185977421272e-06, | |
| "loss": 1.0203, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 5.226381461675579e-06, | |
| "loss": 0.9944, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 5.464052287581699e-06, | |
| "loss": 0.9667, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 5.7012477718360074e-06, | |
| "loss": 0.9407, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 5.938443256090315e-06, | |
| "loss": 0.9194, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 6.176114081996435e-06, | |
| "loss": 0.891, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 6.413784907902556e-06, | |
| "loss": 0.8853, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 6.651455733808676e-06, | |
| "loss": 0.8699, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 6.889126559714796e-06, | |
| "loss": 0.8597, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 7.126797385620916e-06, | |
| "loss": 0.8384, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 7.364468211527036e-06, | |
| "loss": 0.8305, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 7.602139037433156e-06, | |
| "loss": 0.8346, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 7.839809863339275e-06, | |
| "loss": 0.8277, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 8.077005347593585e-06, | |
| "loss": 0.8125, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 8.314676173499705e-06, | |
| "loss": 0.8124, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 8.552346999405825e-06, | |
| "loss": 0.806, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 8.790017825311945e-06, | |
| "loss": 0.8022, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 9.027688651218063e-06, | |
| "loss": 0.7958, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 9.265359477124183e-06, | |
| "loss": 0.7901, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 9.503030303030303e-06, | |
| "loss": 0.7868, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 9.740701128936423e-06, | |
| "loss": 0.7788, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 9.978371954842543e-06, | |
| "loss": 0.7823, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.0216042780748663e-05, | |
| "loss": 0.7767, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 1.0453238265002973e-05, | |
| "loss": 0.7645, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 1.0690909090909091e-05, | |
| "loss": 0.7687, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 1.0928579916815211e-05, | |
| "loss": 0.7664, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 1.1166250742721331e-05, | |
| "loss": 0.7569, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.1403921568627451e-05, | |
| "loss": 0.772, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.1641592394533571e-05, | |
| "loss": 0.7562, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 1.1879263220439691e-05, | |
| "loss": 0.7599, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 1.2116934046345811e-05, | |
| "loss": 0.7524, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 1.2354604872251931e-05, | |
| "loss": 0.7476, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 1.2592275698158052e-05, | |
| "loss": 0.7385, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 1.282947118241236e-05, | |
| "loss": 0.7428, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 1.306714200831848e-05, | |
| "loss": 0.746, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 1.3304337492572788e-05, | |
| "loss": 0.7406, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 1.3542008318478908e-05, | |
| "loss": 0.7264, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 3.45, | |
| "learning_rate": 1.3779679144385028e-05, | |
| "loss": 0.7346, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 1.4017349970291148e-05, | |
| "loss": 0.7266, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 1.4255020796197267e-05, | |
| "loss": 0.7355, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 1.4492691622103387e-05, | |
| "loss": 0.7319, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 1.4730362448009507e-05, | |
| "loss": 0.7304, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 1.4968033273915627e-05, | |
| "loss": 0.7226, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 1.5205228758169936e-05, | |
| "loss": 0.7288, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 1.5442899584076056e-05, | |
| "loss": 0.7214, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 1.5680095068330364e-05, | |
| "loss": 0.7219, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 1.5917765894236484e-05, | |
| "loss": 0.718, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 1.615496137849079e-05, | |
| "loss": 0.7132, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 1.639263220439691e-05, | |
| "loss": 0.7183, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 4.16, | |
| "learning_rate": 1.663030303030303e-05, | |
| "loss": 0.7123, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 4.22, | |
| "learning_rate": 1.686797385620915e-05, | |
| "loss": 0.7013, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 4.28, | |
| "learning_rate": 1.710564468211527e-05, | |
| "loss": 0.7093, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 1.734284016636958e-05, | |
| "loss": 0.704, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 1.75805109922757e-05, | |
| "loss": 0.7047, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 1.781818181818182e-05, | |
| "loss": 0.7041, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 4.52, | |
| "learning_rate": 1.805585264408794e-05, | |
| "loss": 0.7005, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 4.58, | |
| "learning_rate": 1.829352346999406e-05, | |
| "loss": 0.7098, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 1.853119429590018e-05, | |
| "loss": 0.7019, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 1.87688651218063e-05, | |
| "loss": 0.7024, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 1.900653594771242e-05, | |
| "loss": 0.7015, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 1.924420677361854e-05, | |
| "loss": 0.7032, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 1.948187759952466e-05, | |
| "loss": 0.6989, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 4.93, | |
| "learning_rate": 1.971954842543078e-05, | |
| "loss": 0.7023, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 4.99, | |
| "learning_rate": 1.99572192513369e-05, | |
| "loss": 0.7068, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 5.05, | |
| "learning_rate": 1.9978345546973e-05, | |
| "loss": 0.6804, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 5.11, | |
| "learning_rate": 1.9951937677427877e-05, | |
| "loss": 0.6888, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 5.17, | |
| "learning_rate": 1.992552980788275e-05, | |
| "loss": 0.6932, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 5.23, | |
| "learning_rate": 1.9899121938337625e-05, | |
| "loss": 0.6849, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 5.29, | |
| "learning_rate": 1.987276688453159e-05, | |
| "loss": 0.6776, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 5.35, | |
| "learning_rate": 1.9846411830725558e-05, | |
| "loss": 0.6897, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 5.41, | |
| "learning_rate": 1.9820003961180434e-05, | |
| "loss": 0.683, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 5.47, | |
| "learning_rate": 1.979359609163531e-05, | |
| "loss": 0.6834, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 5.53, | |
| "learning_rate": 1.9767188222090182e-05, | |
| "loss": 0.6811, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 5.59, | |
| "learning_rate": 1.9740780352545058e-05, | |
| "loss": 0.6775, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 5.64, | |
| "learning_rate": 1.9714372482999934e-05, | |
| "loss": 0.683, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 5.7, | |
| "learning_rate": 1.968796461345481e-05, | |
| "loss": 0.6799, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 5.76, | |
| "learning_rate": 1.9661556743909686e-05, | |
| "loss": 0.6754, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 5.82, | |
| "learning_rate": 1.9635201690103652e-05, | |
| "loss": 0.684, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 5.88, | |
| "learning_rate": 1.9608793820558528e-05, | |
| "loss": 0.6753, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 5.94, | |
| "learning_rate": 1.9582385951013404e-05, | |
| "loss": 0.6843, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "learning_rate": 1.955597808146828e-05, | |
| "loss": 0.684, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 6.06, | |
| "learning_rate": 1.9529623027662246e-05, | |
| "loss": 0.6578, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 6.12, | |
| "learning_rate": 1.950326797385621e-05, | |
| "loss": 0.6686, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 6.18, | |
| "learning_rate": 1.9476860104311085e-05, | |
| "loss": 0.6696, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 6.24, | |
| "learning_rate": 1.945045223476596e-05, | |
| "loss": 0.6721, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 6.3, | |
| "learning_rate": 1.9424044365220836e-05, | |
| "loss": 0.6645, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 6.36, | |
| "learning_rate": 1.9397689311414803e-05, | |
| "loss": 0.6636, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 6.42, | |
| "learning_rate": 1.937128144186968e-05, | |
| "loss": 0.6644, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 6.48, | |
| "learning_rate": 1.9344873572324555e-05, | |
| "loss": 0.6697, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 6.54, | |
| "learning_rate": 1.931846570277943e-05, | |
| "loss": 0.6738, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 6.6, | |
| "learning_rate": 1.9292057833234306e-05, | |
| "loss": 0.6645, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 6.65, | |
| "learning_rate": 1.9265649963689182e-05, | |
| "loss": 0.6703, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 6.71, | |
| "learning_rate": 1.9239242094144058e-05, | |
| "loss": 0.6667, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 6.77, | |
| "learning_rate": 1.9212834224598934e-05, | |
| "loss": 0.6677, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 6.83, | |
| "learning_rate": 1.9186479170792897e-05, | |
| "loss": 0.6693, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 6.89, | |
| "learning_rate": 1.9160071301247773e-05, | |
| "loss": 0.6614, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 6.95, | |
| "learning_rate": 1.913366343170265e-05, | |
| "loss": 0.6604, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 7.01, | |
| "learning_rate": 1.9107255562157525e-05, | |
| "loss": 0.6591, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 7.07, | |
| "learning_rate": 1.908090050835149e-05, | |
| "loss": 0.6471, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 7.13, | |
| "learning_rate": 1.9054492638806367e-05, | |
| "loss": 0.6539, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 7.19, | |
| "learning_rate": 1.9028084769261243e-05, | |
| "loss": 0.6461, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 7.25, | |
| "learning_rate": 1.900167689971612e-05, | |
| "loss": 0.6514, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 7.31, | |
| "learning_rate": 1.897526903017099e-05, | |
| "loss": 0.6452, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 7.37, | |
| "learning_rate": 1.8948861160625867e-05, | |
| "loss": 0.6513, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 7.43, | |
| "learning_rate": 1.8922453291080743e-05, | |
| "loss": 0.6534, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 7.49, | |
| "learning_rate": 1.889604542153562e-05, | |
| "loss": 0.65, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 7.55, | |
| "learning_rate": 1.8869637551990495e-05, | |
| "loss": 0.6486, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 7.61, | |
| "learning_rate": 1.884328249818446e-05, | |
| "loss": 0.6498, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 7.66, | |
| "learning_rate": 1.8816874628639337e-05, | |
| "loss": 0.6424, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 7.72, | |
| "learning_rate": 1.8790466759094213e-05, | |
| "loss": 0.6492, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 7.78, | |
| "learning_rate": 1.8764058889549086e-05, | |
| "loss": 0.6564, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 7.84, | |
| "learning_rate": 1.8737703835743052e-05, | |
| "loss": 0.6467, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 7.9, | |
| "learning_rate": 1.8711295966197928e-05, | |
| "loss": 0.6461, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 7.96, | |
| "learning_rate": 1.8684888096652804e-05, | |
| "loss": 0.6424, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 8.02, | |
| "learning_rate": 1.865848022710768e-05, | |
| "loss": 0.6487, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 8.08, | |
| "learning_rate": 1.8632072357562556e-05, | |
| "loss": 0.6331, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 8.14, | |
| "learning_rate": 1.8605664488017428e-05, | |
| "loss": 0.6351, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 8.2, | |
| "learning_rate": 1.8579256618472304e-05, | |
| "loss": 0.6298, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 8.26, | |
| "learning_rate": 1.855284874892718e-05, | |
| "loss": 0.6365, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 8.32, | |
| "learning_rate": 1.8526440879382056e-05, | |
| "loss": 0.6371, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 8.38, | |
| "learning_rate": 1.8500085825576025e-05, | |
| "loss": 0.6312, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 8.44, | |
| "learning_rate": 1.847378358750908e-05, | |
| "loss": 0.6438, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 8.5, | |
| "learning_rate": 1.8447375717963954e-05, | |
| "loss": 0.6307, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 8.56, | |
| "learning_rate": 1.842096784841883e-05, | |
| "loss": 0.6377, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 8.62, | |
| "learning_rate": 1.8394559978873706e-05, | |
| "loss": 0.6457, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 8.67, | |
| "learning_rate": 1.8368152109328582e-05, | |
| "loss": 0.6377, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 8.73, | |
| "learning_rate": 1.8341744239783458e-05, | |
| "loss": 0.6394, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 8.79, | |
| "learning_rate": 1.8315389185977424e-05, | |
| "loss": 0.6392, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 8.85, | |
| "learning_rate": 1.82889813164323e-05, | |
| "loss": 0.6424, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 8.91, | |
| "learning_rate": 1.8262573446887173e-05, | |
| "loss": 0.6333, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 8.97, | |
| "learning_rate": 1.823621839308114e-05, | |
| "loss": 0.6422, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 9.03, | |
| "learning_rate": 1.8209810523536015e-05, | |
| "loss": 0.6311, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 9.09, | |
| "learning_rate": 1.818340265399089e-05, | |
| "loss": 0.6238, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 9.15, | |
| "learning_rate": 1.8156994784445767e-05, | |
| "loss": 0.6179, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 9.21, | |
| "learning_rate": 1.813058691490064e-05, | |
| "loss": 0.6344, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 9.27, | |
| "learning_rate": 1.8104179045355515e-05, | |
| "loss": 0.6303, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 9.33, | |
| "learning_rate": 1.807777117581039e-05, | |
| "loss": 0.625, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 9.39, | |
| "learning_rate": 1.8051363306265267e-05, | |
| "loss": 0.6237, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 9.45, | |
| "learning_rate": 1.8024955436720143e-05, | |
| "loss": 0.6332, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 9.51, | |
| "learning_rate": 1.799854756717502e-05, | |
| "loss": 0.6239, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 9.57, | |
| "learning_rate": 1.7972139697629895e-05, | |
| "loss": 0.627, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 9.63, | |
| "learning_rate": 1.794573182808477e-05, | |
| "loss": 0.6202, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 9.69, | |
| "learning_rate": 1.7919376774278737e-05, | |
| "loss": 0.63, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 9.74, | |
| "learning_rate": 1.7892968904733613e-05, | |
| "loss": 0.6205, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 9.8, | |
| "learning_rate": 1.786656103518849e-05, | |
| "loss": 0.6337, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 9.86, | |
| "learning_rate": 1.7840153165643365e-05, | |
| "loss": 0.6234, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 9.92, | |
| "learning_rate": 1.781374529609824e-05, | |
| "loss": 0.6349, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 9.98, | |
| "learning_rate": 1.7787337426553117e-05, | |
| "loss": 0.6336, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 10.04, | |
| "learning_rate": 1.776092955700799e-05, | |
| "loss": 0.6131, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 10.1, | |
| "learning_rate": 1.7734521687462865e-05, | |
| "loss": 0.6177, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 10.16, | |
| "learning_rate": 1.770816663365683e-05, | |
| "loss": 0.6209, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 10.22, | |
| "learning_rate": 1.7681758764111707e-05, | |
| "loss": 0.619, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 10.28, | |
| "learning_rate": 1.7655350894566583e-05, | |
| "loss": 0.6142, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 10.34, | |
| "learning_rate": 1.762894302502146e-05, | |
| "loss": 0.6134, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 10.4, | |
| "learning_rate": 1.7602535155476332e-05, | |
| "loss": 0.6174, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 10.46, | |
| "learning_rate": 1.7576180101670298e-05, | |
| "loss": 0.6203, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 10.52, | |
| "learning_rate": 1.7549772232125174e-05, | |
| "loss": 0.6171, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 10.58, | |
| "learning_rate": 1.752336436258005e-05, | |
| "loss": 0.6186, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 10.64, | |
| "learning_rate": 1.7496956493034926e-05, | |
| "loss": 0.617, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 10.7, | |
| "learning_rate": 1.74705486234898e-05, | |
| "loss": 0.6117, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 10.75, | |
| "learning_rate": 1.7444140753944678e-05, | |
| "loss": 0.6131, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 10.81, | |
| "learning_rate": 1.7417732884399553e-05, | |
| "loss": 0.6187, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 10.87, | |
| "learning_rate": 1.739132501485443e-05, | |
| "loss": 0.6215, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 10.93, | |
| "learning_rate": 1.7364969961048396e-05, | |
| "loss": 0.6237, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 10.99, | |
| "learning_rate": 1.733856209150327e-05, | |
| "loss": 0.6269, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 11.05, | |
| "learning_rate": 1.7312207037697234e-05, | |
| "loss": 0.6065, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 11.11, | |
| "learning_rate": 1.728579916815211e-05, | |
| "loss": 0.603, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 11.17, | |
| "learning_rate": 1.7259391298606986e-05, | |
| "loss": 0.6037, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 11.23, | |
| "learning_rate": 1.7232983429061862e-05, | |
| "loss": 0.6081, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 11.29, | |
| "learning_rate": 1.7206575559516738e-05, | |
| "loss": 0.602, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 11.35, | |
| "learning_rate": 1.718016768997161e-05, | |
| "loss": 0.6132, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 11.41, | |
| "learning_rate": 1.7153759820426487e-05, | |
| "loss": 0.6049, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 11.47, | |
| "learning_rate": 1.7127351950881362e-05, | |
| "loss": 0.6088, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 11.53, | |
| "learning_rate": 1.710094408133624e-05, | |
| "loss": 0.6063, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 11.59, | |
| "learning_rate": 1.7074589027530205e-05, | |
| "loss": 0.6086, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 11.65, | |
| "learning_rate": 1.704818115798508e-05, | |
| "loss": 0.6102, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 11.71, | |
| "learning_rate": 1.7021826104179047e-05, | |
| "loss": 0.6068, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 11.76, | |
| "learning_rate": 1.6995418234633923e-05, | |
| "loss": 0.6122, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 11.82, | |
| "learning_rate": 1.69690103650888e-05, | |
| "loss": 0.6119, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 11.88, | |
| "learning_rate": 1.6942602495543675e-05, | |
| "loss": 0.6058, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 11.94, | |
| "learning_rate": 1.691619462599855e-05, | |
| "loss": 0.6158, | |
| "step": 100500 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "learning_rate": 1.6889786756453423e-05, | |
| "loss": 0.6219, | |
| "step": 101000 | |
| }, | |
| { | |
| "epoch": 12.06, | |
| "learning_rate": 1.68633788869083e-05, | |
| "loss": 0.5941, | |
| "step": 101500 | |
| }, | |
| { | |
| "epoch": 12.12, | |
| "learning_rate": 1.6836971017363175e-05, | |
| "loss": 0.5933, | |
| "step": 102000 | |
| }, | |
| { | |
| "epoch": 12.18, | |
| "learning_rate": 1.681056314781805e-05, | |
| "loss": 0.6045, | |
| "step": 102500 | |
| }, | |
| { | |
| "epoch": 12.24, | |
| "learning_rate": 1.6784155278272927e-05, | |
| "loss": 0.5995, | |
| "step": 103000 | |
| }, | |
| { | |
| "epoch": 12.3, | |
| "learning_rate": 1.6757747408727803e-05, | |
| "loss": 0.5976, | |
| "step": 103500 | |
| }, | |
| { | |
| "epoch": 12.36, | |
| "learning_rate": 1.673139235492177e-05, | |
| "loss": 0.6016, | |
| "step": 104000 | |
| }, | |
| { | |
| "epoch": 12.42, | |
| "learning_rate": 1.670498448537664e-05, | |
| "loss": 0.6029, | |
| "step": 104500 | |
| }, | |
| { | |
| "epoch": 12.48, | |
| "learning_rate": 1.6678576615831517e-05, | |
| "loss": 0.6008, | |
| "step": 105000 | |
| }, | |
| { | |
| "epoch": 12.54, | |
| "learning_rate": 1.6652221562025487e-05, | |
| "loss": 0.5978, | |
| "step": 105500 | |
| }, | |
| { | |
| "epoch": 12.6, | |
| "learning_rate": 1.6625813692480363e-05, | |
| "loss": 0.6029, | |
| "step": 106000 | |
| }, | |
| { | |
| "epoch": 12.66, | |
| "learning_rate": 1.6599405822935235e-05, | |
| "loss": 0.6088, | |
| "step": 106500 | |
| }, | |
| { | |
| "epoch": 12.72, | |
| "learning_rate": 1.657299795339011e-05, | |
| "loss": 0.6071, | |
| "step": 107000 | |
| }, | |
| { | |
| "epoch": 12.77, | |
| "learning_rate": 1.6546590083844987e-05, | |
| "loss": 0.6025, | |
| "step": 107500 | |
| }, | |
| { | |
| "epoch": 12.83, | |
| "learning_rate": 1.6520182214299863e-05, | |
| "loss": 0.5966, | |
| "step": 108000 | |
| }, | |
| { | |
| "epoch": 12.89, | |
| "learning_rate": 1.649377434475474e-05, | |
| "loss": 0.6058, | |
| "step": 108500 | |
| }, | |
| { | |
| "epoch": 12.95, | |
| "learning_rate": 1.6467419290948702e-05, | |
| "loss": 0.6109, | |
| "step": 109000 | |
| }, | |
| { | |
| "epoch": 13.01, | |
| "learning_rate": 1.6441011421403578e-05, | |
| "loss": 0.6024, | |
| "step": 109500 | |
| }, | |
| { | |
| "epoch": 13.07, | |
| "learning_rate": 1.6414603551858454e-05, | |
| "loss": 0.5963, | |
| "step": 110000 | |
| }, | |
| { | |
| "epoch": 13.13, | |
| "learning_rate": 1.638819568231333e-05, | |
| "loss": 0.5861, | |
| "step": 110500 | |
| }, | |
| { | |
| "epoch": 13.19, | |
| "learning_rate": 1.6361787812768206e-05, | |
| "loss": 0.5939, | |
| "step": 111000 | |
| }, | |
| { | |
| "epoch": 13.25, | |
| "learning_rate": 1.633537994322308e-05, | |
| "loss": 0.5859, | |
| "step": 111500 | |
| }, | |
| { | |
| "epoch": 13.31, | |
| "learning_rate": 1.6308972073677957e-05, | |
| "loss": 0.5916, | |
| "step": 112000 | |
| }, | |
| { | |
| "epoch": 13.37, | |
| "learning_rate": 1.6282564204132833e-05, | |
| "loss": 0.5938, | |
| "step": 112500 | |
| }, | |
| { | |
| "epoch": 13.43, | |
| "learning_rate": 1.62562091503268e-05, | |
| "loss": 0.5925, | |
| "step": 113000 | |
| }, | |
| { | |
| "epoch": 13.49, | |
| "learning_rate": 1.6229801280781675e-05, | |
| "loss": 0.5945, | |
| "step": 113500 | |
| }, | |
| { | |
| "epoch": 13.55, | |
| "learning_rate": 1.620339341123655e-05, | |
| "loss": 0.5999, | |
| "step": 114000 | |
| }, | |
| { | |
| "epoch": 13.61, | |
| "learning_rate": 1.6176985541691427e-05, | |
| "loss": 0.5904, | |
| "step": 114500 | |
| }, | |
| { | |
| "epoch": 13.67, | |
| "learning_rate": 1.615063048788539e-05, | |
| "loss": 0.5988, | |
| "step": 115000 | |
| }, | |
| { | |
| "epoch": 13.73, | |
| "learning_rate": 1.6124222618340266e-05, | |
| "loss": 0.5963, | |
| "step": 115500 | |
| }, | |
| { | |
| "epoch": 13.78, | |
| "learning_rate": 1.6097814748795142e-05, | |
| "loss": 0.592, | |
| "step": 116000 | |
| }, | |
| { | |
| "epoch": 13.84, | |
| "learning_rate": 1.6071406879250018e-05, | |
| "loss": 0.5954, | |
| "step": 116500 | |
| }, | |
| { | |
| "epoch": 13.9, | |
| "learning_rate": 1.6044999009704894e-05, | |
| "loss": 0.5966, | |
| "step": 117000 | |
| }, | |
| { | |
| "epoch": 13.96, | |
| "learning_rate": 1.601859114015977e-05, | |
| "loss": 0.6015, | |
| "step": 117500 | |
| }, | |
| { | |
| "epoch": 14.02, | |
| "learning_rate": 1.5992183270614646e-05, | |
| "loss": 0.5891, | |
| "step": 118000 | |
| }, | |
| { | |
| "epoch": 14.08, | |
| "learning_rate": 1.596577540106952e-05, | |
| "loss": 0.5819, | |
| "step": 118500 | |
| }, | |
| { | |
| "epoch": 14.14, | |
| "learning_rate": 1.5939367531524394e-05, | |
| "loss": 0.5883, | |
| "step": 119000 | |
| }, | |
| { | |
| "epoch": 14.2, | |
| "learning_rate": 1.591301247771836e-05, | |
| "loss": 0.5881, | |
| "step": 119500 | |
| }, | |
| { | |
| "epoch": 14.26, | |
| "learning_rate": 1.5886604608173236e-05, | |
| "loss": 0.584, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 14.32, | |
| "learning_rate": 1.5860196738628112e-05, | |
| "loss": 0.5888, | |
| "step": 120500 | |
| }, | |
| { | |
| "epoch": 14.38, | |
| "learning_rate": 1.5833788869082988e-05, | |
| "loss": 0.5911, | |
| "step": 121000 | |
| }, | |
| { | |
| "epoch": 14.44, | |
| "learning_rate": 1.5807380999537864e-05, | |
| "loss": 0.5846, | |
| "step": 121500 | |
| }, | |
| { | |
| "epoch": 14.5, | |
| "learning_rate": 1.5780973129992737e-05, | |
| "loss": 0.5836, | |
| "step": 122000 | |
| }, | |
| { | |
| "epoch": 14.56, | |
| "learning_rate": 1.5754565260447613e-05, | |
| "loss": 0.5838, | |
| "step": 122500 | |
| }, | |
| { | |
| "epoch": 14.62, | |
| "learning_rate": 1.5728210206641582e-05, | |
| "loss": 0.5883, | |
| "step": 123000 | |
| }, | |
| { | |
| "epoch": 14.68, | |
| "learning_rate": 1.5701802337096458e-05, | |
| "loss": 0.5875, | |
| "step": 123500 | |
| }, | |
| { | |
| "epoch": 14.74, | |
| "learning_rate": 1.5675394467551334e-05, | |
| "loss": 0.5901, | |
| "step": 124000 | |
| }, | |
| { | |
| "epoch": 14.79, | |
| "learning_rate": 1.5648986598006207e-05, | |
| "loss": 0.5881, | |
| "step": 124500 | |
| }, | |
| { | |
| "epoch": 14.85, | |
| "learning_rate": 1.5622631544200173e-05, | |
| "loss": 0.5951, | |
| "step": 125000 | |
| }, | |
| { | |
| "epoch": 14.91, | |
| "learning_rate": 1.559622367465505e-05, | |
| "loss": 0.5898, | |
| "step": 125500 | |
| }, | |
| { | |
| "epoch": 14.97, | |
| "learning_rate": 1.5569815805109925e-05, | |
| "loss": 0.5861, | |
| "step": 126000 | |
| }, | |
| { | |
| "epoch": 15.03, | |
| "learning_rate": 1.55434079355648e-05, | |
| "loss": 0.5754, | |
| "step": 126500 | |
| }, | |
| { | |
| "epoch": 15.09, | |
| "learning_rate": 1.5517000066019676e-05, | |
| "loss": 0.5751, | |
| "step": 127000 | |
| }, | |
| { | |
| "epoch": 15.15, | |
| "learning_rate": 1.549064501221364e-05, | |
| "loss": 0.5814, | |
| "step": 127500 | |
| }, | |
| { | |
| "epoch": 15.21, | |
| "learning_rate": 1.5464237142668515e-05, | |
| "loss": 0.5853, | |
| "step": 128000 | |
| }, | |
| { | |
| "epoch": 15.27, | |
| "learning_rate": 1.543782927312339e-05, | |
| "loss": 0.5832, | |
| "step": 128500 | |
| }, | |
| { | |
| "epoch": 15.33, | |
| "learning_rate": 1.5411421403578267e-05, | |
| "loss": 0.5831, | |
| "step": 129000 | |
| }, | |
| { | |
| "epoch": 15.39, | |
| "learning_rate": 1.5385013534033143e-05, | |
| "loss": 0.583, | |
| "step": 129500 | |
| }, | |
| { | |
| "epoch": 15.45, | |
| "learning_rate": 1.535860566448802e-05, | |
| "loss": 0.5708, | |
| "step": 130000 | |
| }, | |
| { | |
| "epoch": 15.51, | |
| "learning_rate": 1.5332250610681985e-05, | |
| "loss": 0.5816, | |
| "step": 130500 | |
| }, | |
| { | |
| "epoch": 15.57, | |
| "learning_rate": 1.530584274113686e-05, | |
| "loss": 0.5845, | |
| "step": 131000 | |
| }, | |
| { | |
| "epoch": 15.63, | |
| "learning_rate": 1.5279434871591737e-05, | |
| "loss": 0.5821, | |
| "step": 131500 | |
| }, | |
| { | |
| "epoch": 15.69, | |
| "learning_rate": 1.5253027002046611e-05, | |
| "loss": 0.5835, | |
| "step": 132000 | |
| }, | |
| { | |
| "epoch": 15.75, | |
| "learning_rate": 1.5226619132501485e-05, | |
| "loss": 0.5796, | |
| "step": 132500 | |
| }, | |
| { | |
| "epoch": 15.81, | |
| "learning_rate": 1.5200211262956361e-05, | |
| "loss": 0.5825, | |
| "step": 133000 | |
| }, | |
| { | |
| "epoch": 15.86, | |
| "learning_rate": 1.5173803393411237e-05, | |
| "loss": 0.5854, | |
| "step": 133500 | |
| }, | |
| { | |
| "epoch": 15.92, | |
| "learning_rate": 1.5147395523866113e-05, | |
| "loss": 0.5813, | |
| "step": 134000 | |
| }, | |
| { | |
| "epoch": 15.98, | |
| "learning_rate": 1.5121093285799168e-05, | |
| "loss": 0.584, | |
| "step": 134500 | |
| }, | |
| { | |
| "epoch": 16.04, | |
| "learning_rate": 1.5094685416254044e-05, | |
| "loss": 0.5734, | |
| "step": 135000 | |
| }, | |
| { | |
| "epoch": 16.1, | |
| "learning_rate": 1.506827754670892e-05, | |
| "loss": 0.5647, | |
| "step": 135500 | |
| }, | |
| { | |
| "epoch": 16.16, | |
| "learning_rate": 1.5041869677163796e-05, | |
| "loss": 0.5729, | |
| "step": 136000 | |
| }, | |
| { | |
| "epoch": 16.22, | |
| "learning_rate": 1.501551462335776e-05, | |
| "loss": 0.5692, | |
| "step": 136500 | |
| }, | |
| { | |
| "epoch": 16.28, | |
| "learning_rate": 1.4989159569551728e-05, | |
| "loss": 0.5703, | |
| "step": 137000 | |
| }, | |
| { | |
| "epoch": 16.34, | |
| "learning_rate": 1.4962751700006604e-05, | |
| "loss": 0.5798, | |
| "step": 137500 | |
| }, | |
| { | |
| "epoch": 16.4, | |
| "learning_rate": 1.493634383046148e-05, | |
| "loss": 0.5785, | |
| "step": 138000 | |
| }, | |
| { | |
| "epoch": 16.46, | |
| "learning_rate": 1.4909935960916356e-05, | |
| "loss": 0.57, | |
| "step": 138500 | |
| }, | |
| { | |
| "epoch": 16.52, | |
| "learning_rate": 1.488358090711032e-05, | |
| "loss": 0.5747, | |
| "step": 139000 | |
| }, | |
| { | |
| "epoch": 16.58, | |
| "learning_rate": 1.4857173037565196e-05, | |
| "loss": 0.5761, | |
| "step": 139500 | |
| }, | |
| { | |
| "epoch": 16.64, | |
| "learning_rate": 1.483076516802007e-05, | |
| "loss": 0.5756, | |
| "step": 140000 | |
| }, | |
| { | |
| "epoch": 16.7, | |
| "learning_rate": 1.4804357298474947e-05, | |
| "loss": 0.5727, | |
| "step": 140500 | |
| }, | |
| { | |
| "epoch": 16.76, | |
| "learning_rate": 1.4777949428929823e-05, | |
| "loss": 0.5826, | |
| "step": 141000 | |
| }, | |
| { | |
| "epoch": 16.82, | |
| "learning_rate": 1.4751541559384698e-05, | |
| "loss": 0.5862, | |
| "step": 141500 | |
| }, | |
| { | |
| "epoch": 16.87, | |
| "learning_rate": 1.4725133689839573e-05, | |
| "loss": 0.5752, | |
| "step": 142000 | |
| }, | |
| { | |
| "epoch": 16.93, | |
| "learning_rate": 1.4698725820294449e-05, | |
| "loss": 0.583, | |
| "step": 142500 | |
| }, | |
| { | |
| "epoch": 16.99, | |
| "learning_rate": 1.4672317950749325e-05, | |
| "loss": 0.5805, | |
| "step": 143000 | |
| }, | |
| { | |
| "epoch": 17.05, | |
| "learning_rate": 1.464596289694329e-05, | |
| "loss": 0.566, | |
| "step": 143500 | |
| }, | |
| { | |
| "epoch": 17.11, | |
| "learning_rate": 1.4619555027398167e-05, | |
| "loss": 0.5645, | |
| "step": 144000 | |
| }, | |
| { | |
| "epoch": 17.17, | |
| "learning_rate": 1.4593199973592131e-05, | |
| "loss": 0.5645, | |
| "step": 144500 | |
| }, | |
| { | |
| "epoch": 17.23, | |
| "learning_rate": 1.4566792104047007e-05, | |
| "loss": 0.5691, | |
| "step": 145000 | |
| }, | |
| { | |
| "epoch": 17.29, | |
| "learning_rate": 1.4540384234501883e-05, | |
| "loss": 0.5693, | |
| "step": 145500 | |
| }, | |
| { | |
| "epoch": 17.35, | |
| "learning_rate": 1.4513976364956759e-05, | |
| "loss": 0.5731, | |
| "step": 146000 | |
| }, | |
| { | |
| "epoch": 17.41, | |
| "learning_rate": 1.4487568495411635e-05, | |
| "loss": 0.5658, | |
| "step": 146500 | |
| }, | |
| { | |
| "epoch": 17.47, | |
| "learning_rate": 1.446116062586651e-05, | |
| "loss": 0.5761, | |
| "step": 147000 | |
| }, | |
| { | |
| "epoch": 17.53, | |
| "learning_rate": 1.4434752756321383e-05, | |
| "loss": 0.5687, | |
| "step": 147500 | |
| }, | |
| { | |
| "epoch": 17.59, | |
| "learning_rate": 1.440834488677626e-05, | |
| "loss": 0.5724, | |
| "step": 148000 | |
| }, | |
| { | |
| "epoch": 17.65, | |
| "learning_rate": 1.4381937017231135e-05, | |
| "loss": 0.5683, | |
| "step": 148500 | |
| }, | |
| { | |
| "epoch": 17.71, | |
| "learning_rate": 1.4355581963425103e-05, | |
| "loss": 0.5772, | |
| "step": 149000 | |
| }, | |
| { | |
| "epoch": 17.77, | |
| "learning_rate": 1.4329174093879979e-05, | |
| "loss": 0.5669, | |
| "step": 149500 | |
| }, | |
| { | |
| "epoch": 17.83, | |
| "learning_rate": 1.4302819040073944e-05, | |
| "loss": 0.5678, | |
| "step": 150000 | |
| }, | |
| { | |
| "epoch": 17.88, | |
| "learning_rate": 1.427646398626791e-05, | |
| "loss": 0.5703, | |
| "step": 150500 | |
| }, | |
| { | |
| "epoch": 17.94, | |
| "learning_rate": 1.4250056116722784e-05, | |
| "loss": 0.5687, | |
| "step": 151000 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "learning_rate": 1.422364824717766e-05, | |
| "loss": 0.5656, | |
| "step": 151500 | |
| }, | |
| { | |
| "epoch": 18.06, | |
| "learning_rate": 1.4197240377632536e-05, | |
| "loss": 0.5563, | |
| "step": 152000 | |
| }, | |
| { | |
| "epoch": 18.12, | |
| "learning_rate": 1.4170885323826502e-05, | |
| "loss": 0.5584, | |
| "step": 152500 | |
| }, | |
| { | |
| "epoch": 18.18, | |
| "learning_rate": 1.4144477454281378e-05, | |
| "loss": 0.5699, | |
| "step": 153000 | |
| }, | |
| { | |
| "epoch": 18.24, | |
| "learning_rate": 1.4118069584736254e-05, | |
| "loss": 0.5573, | |
| "step": 153500 | |
| }, | |
| { | |
| "epoch": 18.3, | |
| "learning_rate": 1.4091661715191126e-05, | |
| "loss": 0.5646, | |
| "step": 154000 | |
| }, | |
| { | |
| "epoch": 18.36, | |
| "learning_rate": 1.4065306661385094e-05, | |
| "loss": 0.5608, | |
| "step": 154500 | |
| }, | |
| { | |
| "epoch": 18.42, | |
| "learning_rate": 1.4038951607579059e-05, | |
| "loss": 0.566, | |
| "step": 155000 | |
| }, | |
| { | |
| "epoch": 18.48, | |
| "learning_rate": 1.4012543738033935e-05, | |
| "loss": 0.5709, | |
| "step": 155500 | |
| }, | |
| { | |
| "epoch": 18.54, | |
| "learning_rate": 1.398613586848881e-05, | |
| "loss": 0.5665, | |
| "step": 156000 | |
| }, | |
| { | |
| "epoch": 18.6, | |
| "learning_rate": 1.3959727998943687e-05, | |
| "loss": 0.5642, | |
| "step": 156500 | |
| }, | |
| { | |
| "epoch": 18.66, | |
| "learning_rate": 1.3933320129398563e-05, | |
| "loss": 0.5647, | |
| "step": 157000 | |
| }, | |
| { | |
| "epoch": 18.72, | |
| "learning_rate": 1.3906912259853439e-05, | |
| "loss": 0.5716, | |
| "step": 157500 | |
| }, | |
| { | |
| "epoch": 18.78, | |
| "learning_rate": 1.3880504390308314e-05, | |
| "loss": 0.5673, | |
| "step": 158000 | |
| }, | |
| { | |
| "epoch": 18.84, | |
| "learning_rate": 1.3854096520763189e-05, | |
| "loss": 0.5671, | |
| "step": 158500 | |
| }, | |
| { | |
| "epoch": 18.89, | |
| "learning_rate": 1.3827688651218063e-05, | |
| "loss": 0.5657, | |
| "step": 159000 | |
| }, | |
| { | |
| "epoch": 18.95, | |
| "learning_rate": 1.3801280781672939e-05, | |
| "loss": 0.5629, | |
| "step": 159500 | |
| }, | |
| { | |
| "epoch": 19.01, | |
| "learning_rate": 1.3774872912127815e-05, | |
| "loss": 0.5674, | |
| "step": 160000 | |
| }, | |
| { | |
| "epoch": 19.07, | |
| "learning_rate": 1.374846504258269e-05, | |
| "loss": 0.5609, | |
| "step": 160500 | |
| }, | |
| { | |
| "epoch": 19.13, | |
| "learning_rate": 1.3722057173037567e-05, | |
| "loss": 0.5642, | |
| "step": 161000 | |
| }, | |
| { | |
| "epoch": 19.19, | |
| "learning_rate": 1.3695649303492443e-05, | |
| "loss": 0.5545, | |
| "step": 161500 | |
| }, | |
| { | |
| "epoch": 19.25, | |
| "learning_rate": 1.3669241433947318e-05, | |
| "loss": 0.5548, | |
| "step": 162000 | |
| }, | |
| { | |
| "epoch": 19.31, | |
| "learning_rate": 1.3642833564402194e-05, | |
| "loss": 0.5523, | |
| "step": 162500 | |
| }, | |
| { | |
| "epoch": 19.37, | |
| "learning_rate": 1.3616425694857069e-05, | |
| "loss": 0.5602, | |
| "step": 163000 | |
| }, | |
| { | |
| "epoch": 19.43, | |
| "learning_rate": 1.3590017825311943e-05, | |
| "loss": 0.5627, | |
| "step": 163500 | |
| }, | |
| { | |
| "epoch": 19.49, | |
| "learning_rate": 1.3563609955766819e-05, | |
| "loss": 0.5549, | |
| "step": 164000 | |
| }, | |
| { | |
| "epoch": 19.55, | |
| "learning_rate": 1.3537202086221695e-05, | |
| "loss": 0.5606, | |
| "step": 164500 | |
| }, | |
| { | |
| "epoch": 19.61, | |
| "learning_rate": 1.351079421667657e-05, | |
| "loss": 0.5672, | |
| "step": 165000 | |
| }, | |
| { | |
| "epoch": 19.67, | |
| "learning_rate": 1.3484386347131447e-05, | |
| "loss": 0.5658, | |
| "step": 165500 | |
| }, | |
| { | |
| "epoch": 19.73, | |
| "learning_rate": 1.3457978477586322e-05, | |
| "loss": 0.5628, | |
| "step": 166000 | |
| }, | |
| { | |
| "epoch": 19.79, | |
| "learning_rate": 1.3431570608041198e-05, | |
| "loss": 0.5612, | |
| "step": 166500 | |
| }, | |
| { | |
| "epoch": 19.85, | |
| "learning_rate": 1.3405162738496074e-05, | |
| "loss": 0.5643, | |
| "step": 167000 | |
| }, | |
| { | |
| "epoch": 19.9, | |
| "learning_rate": 1.337875486895095e-05, | |
| "loss": 0.568, | |
| "step": 167500 | |
| }, | |
| { | |
| "epoch": 19.96, | |
| "learning_rate": 1.3352346999405823e-05, | |
| "loss": 0.563, | |
| "step": 168000 | |
| }, | |
| { | |
| "epoch": 20.02, | |
| "learning_rate": 1.332599194559979e-05, | |
| "loss": 0.5639, | |
| "step": 168500 | |
| }, | |
| { | |
| "epoch": 20.08, | |
| "learning_rate": 1.3299584076054665e-05, | |
| "loss": 0.5526, | |
| "step": 169000 | |
| }, | |
| { | |
| "epoch": 20.14, | |
| "learning_rate": 1.327317620650954e-05, | |
| "loss": 0.5549, | |
| "step": 169500 | |
| }, | |
| { | |
| "epoch": 20.2, | |
| "learning_rate": 1.3246768336964417e-05, | |
| "loss": 0.5477, | |
| "step": 170000 | |
| }, | |
| { | |
| "epoch": 20.26, | |
| "learning_rate": 1.3220360467419293e-05, | |
| "loss": 0.5534, | |
| "step": 170500 | |
| }, | |
| { | |
| "epoch": 20.32, | |
| "learning_rate": 1.3194005413613257e-05, | |
| "loss": 0.5555, | |
| "step": 171000 | |
| }, | |
| { | |
| "epoch": 20.38, | |
| "learning_rate": 1.3167597544068133e-05, | |
| "loss": 0.5646, | |
| "step": 171500 | |
| }, | |
| { | |
| "epoch": 20.44, | |
| "learning_rate": 1.3141189674523009e-05, | |
| "loss": 0.5538, | |
| "step": 172000 | |
| }, | |
| { | |
| "epoch": 20.5, | |
| "learning_rate": 1.3114781804977885e-05, | |
| "loss": 0.5567, | |
| "step": 172500 | |
| }, | |
| { | |
| "epoch": 20.56, | |
| "learning_rate": 1.3088373935432761e-05, | |
| "loss": 0.5553, | |
| "step": 173000 | |
| }, | |
| { | |
| "epoch": 20.62, | |
| "learning_rate": 1.3061966065887635e-05, | |
| "loss": 0.5573, | |
| "step": 173500 | |
| }, | |
| { | |
| "epoch": 20.68, | |
| "learning_rate": 1.3035611012081601e-05, | |
| "loss": 0.5624, | |
| "step": 174000 | |
| }, | |
| { | |
| "epoch": 20.74, | |
| "learning_rate": 1.3009255958275566e-05, | |
| "loss": 0.5553, | |
| "step": 174500 | |
| }, | |
| { | |
| "epoch": 20.8, | |
| "learning_rate": 1.2982848088730442e-05, | |
| "loss": 0.57, | |
| "step": 175000 | |
| }, | |
| { | |
| "epoch": 20.86, | |
| "learning_rate": 1.2956440219185318e-05, | |
| "loss": 0.5502, | |
| "step": 175500 | |
| }, | |
| { | |
| "epoch": 20.91, | |
| "learning_rate": 1.2930032349640194e-05, | |
| "loss": 0.5544, | |
| "step": 176000 | |
| }, | |
| { | |
| "epoch": 20.97, | |
| "learning_rate": 1.290367729583416e-05, | |
| "loss": 0.5627, | |
| "step": 176500 | |
| }, | |
| { | |
| "epoch": 21.03, | |
| "learning_rate": 1.2877269426289034e-05, | |
| "loss": 0.5574, | |
| "step": 177000 | |
| }, | |
| { | |
| "epoch": 21.09, | |
| "learning_rate": 1.285086155674391e-05, | |
| "loss": 0.5469, | |
| "step": 177500 | |
| }, | |
| { | |
| "epoch": 21.15, | |
| "learning_rate": 1.2824453687198786e-05, | |
| "loss": 0.5495, | |
| "step": 178000 | |
| }, | |
| { | |
| "epoch": 21.21, | |
| "learning_rate": 1.2798045817653662e-05, | |
| "loss": 0.5523, | |
| "step": 178500 | |
| }, | |
| { | |
| "epoch": 21.27, | |
| "learning_rate": 1.2771690763847628e-05, | |
| "loss": 0.5467, | |
| "step": 179000 | |
| }, | |
| { | |
| "epoch": 21.33, | |
| "learning_rate": 1.2745282894302504e-05, | |
| "loss": 0.5487, | |
| "step": 179500 | |
| }, | |
| { | |
| "epoch": 21.39, | |
| "learning_rate": 1.2718875024757378e-05, | |
| "loss": 0.546, | |
| "step": 180000 | |
| }, | |
| { | |
| "epoch": 21.45, | |
| "learning_rate": 1.2692467155212254e-05, | |
| "loss": 0.5461, | |
| "step": 180500 | |
| }, | |
| { | |
| "epoch": 21.51, | |
| "learning_rate": 1.266605928566713e-05, | |
| "loss": 0.554, | |
| "step": 181000 | |
| }, | |
| { | |
| "epoch": 21.57, | |
| "learning_rate": 1.2639704231861096e-05, | |
| "loss": 0.5493, | |
| "step": 181500 | |
| }, | |
| { | |
| "epoch": 21.63, | |
| "learning_rate": 1.2613296362315972e-05, | |
| "loss": 0.5548, | |
| "step": 182000 | |
| }, | |
| { | |
| "epoch": 21.69, | |
| "learning_rate": 1.2586888492770846e-05, | |
| "loss": 0.5503, | |
| "step": 182500 | |
| }, | |
| { | |
| "epoch": 21.75, | |
| "learning_rate": 1.256048062322572e-05, | |
| "loss": 0.5604, | |
| "step": 183000 | |
| }, | |
| { | |
| "epoch": 21.81, | |
| "learning_rate": 1.2534072753680597e-05, | |
| "loss": 0.5622, | |
| "step": 183500 | |
| }, | |
| { | |
| "epoch": 21.87, | |
| "learning_rate": 1.2507664884135473e-05, | |
| "loss": 0.549, | |
| "step": 184000 | |
| }, | |
| { | |
| "epoch": 21.93, | |
| "learning_rate": 1.2481257014590348e-05, | |
| "loss": 0.5535, | |
| "step": 184500 | |
| }, | |
| { | |
| "epoch": 21.98, | |
| "learning_rate": 1.2454849145045224e-05, | |
| "loss": 0.5572, | |
| "step": 185000 | |
| }, | |
| { | |
| "epoch": 22.04, | |
| "learning_rate": 1.24284412755001e-05, | |
| "loss": 0.5508, | |
| "step": 185500 | |
| }, | |
| { | |
| "epoch": 22.1, | |
| "learning_rate": 1.2402033405954976e-05, | |
| "loss": 0.5454, | |
| "step": 186000 | |
| }, | |
| { | |
| "epoch": 22.16, | |
| "learning_rate": 1.237567835214894e-05, | |
| "loss": 0.551, | |
| "step": 186500 | |
| }, | |
| { | |
| "epoch": 22.22, | |
| "learning_rate": 1.2349270482603817e-05, | |
| "loss": 0.5454, | |
| "step": 187000 | |
| }, | |
| { | |
| "epoch": 22.28, | |
| "learning_rate": 1.2322862613058693e-05, | |
| "loss": 0.5445, | |
| "step": 187500 | |
| }, | |
| { | |
| "epoch": 22.34, | |
| "learning_rate": 1.2296454743513569e-05, | |
| "loss": 0.5493, | |
| "step": 188000 | |
| }, | |
| { | |
| "epoch": 22.4, | |
| "learning_rate": 1.2270046873968444e-05, | |
| "loss": 0.5429, | |
| "step": 188500 | |
| }, | |
| { | |
| "epoch": 22.46, | |
| "learning_rate": 1.224363900442332e-05, | |
| "loss": 0.5507, | |
| "step": 189000 | |
| }, | |
| { | |
| "epoch": 22.52, | |
| "learning_rate": 1.2217231134878196e-05, | |
| "loss": 0.5439, | |
| "step": 189500 | |
| }, | |
| { | |
| "epoch": 22.58, | |
| "learning_rate": 1.2190823265333069e-05, | |
| "loss": 0.5484, | |
| "step": 190000 | |
| }, | |
| { | |
| "epoch": 22.64, | |
| "learning_rate": 1.2164468211527037e-05, | |
| "loss": 0.5467, | |
| "step": 190500 | |
| }, | |
| { | |
| "epoch": 22.7, | |
| "learning_rate": 1.2138113157721001e-05, | |
| "loss": 0.5501, | |
| "step": 191000 | |
| }, | |
| { | |
| "epoch": 22.76, | |
| "learning_rate": 1.2111705288175877e-05, | |
| "loss": 0.5495, | |
| "step": 191500 | |
| }, | |
| { | |
| "epoch": 22.82, | |
| "learning_rate": 1.2085297418630753e-05, | |
| "loss": 0.552, | |
| "step": 192000 | |
| }, | |
| { | |
| "epoch": 22.88, | |
| "learning_rate": 1.205894236482472e-05, | |
| "loss": 0.548, | |
| "step": 192500 | |
| }, | |
| { | |
| "epoch": 22.94, | |
| "learning_rate": 1.2032534495279595e-05, | |
| "loss": 0.5501, | |
| "step": 193000 | |
| }, | |
| { | |
| "epoch": 22.99, | |
| "learning_rate": 1.200617944147356e-05, | |
| "loss": 0.5499, | |
| "step": 193500 | |
| }, | |
| { | |
| "epoch": 23.05, | |
| "learning_rate": 1.1979771571928436e-05, | |
| "loss": 0.5376, | |
| "step": 194000 | |
| }, | |
| { | |
| "epoch": 23.11, | |
| "learning_rate": 1.1953363702383312e-05, | |
| "loss": 0.5413, | |
| "step": 194500 | |
| }, | |
| { | |
| "epoch": 23.17, | |
| "learning_rate": 1.1926955832838188e-05, | |
| "loss": 0.5504, | |
| "step": 195000 | |
| }, | |
| { | |
| "epoch": 23.23, | |
| "learning_rate": 1.1900547963293064e-05, | |
| "loss": 0.5448, | |
| "step": 195500 | |
| }, | |
| { | |
| "epoch": 23.29, | |
| "learning_rate": 1.1874140093747936e-05, | |
| "loss": 0.553, | |
| "step": 196000 | |
| }, | |
| { | |
| "epoch": 23.35, | |
| "learning_rate": 1.1847732224202812e-05, | |
| "loss": 0.5535, | |
| "step": 196500 | |
| }, | |
| { | |
| "epoch": 23.41, | |
| "learning_rate": 1.1821324354657688e-05, | |
| "loss": 0.5449, | |
| "step": 197000 | |
| }, | |
| { | |
| "epoch": 23.47, | |
| "learning_rate": 1.1794916485112564e-05, | |
| "loss": 0.5452, | |
| "step": 197500 | |
| }, | |
| { | |
| "epoch": 23.53, | |
| "learning_rate": 1.176850861556744e-05, | |
| "loss": 0.5496, | |
| "step": 198000 | |
| }, | |
| { | |
| "epoch": 23.59, | |
| "learning_rate": 1.1742100746022316e-05, | |
| "loss": 0.5528, | |
| "step": 198500 | |
| }, | |
| { | |
| "epoch": 23.65, | |
| "learning_rate": 1.1715692876477192e-05, | |
| "loss": 0.5388, | |
| "step": 199000 | |
| }, | |
| { | |
| "epoch": 23.71, | |
| "learning_rate": 1.1689337822671156e-05, | |
| "loss": 0.5412, | |
| "step": 199500 | |
| }, | |
| { | |
| "epoch": 23.77, | |
| "learning_rate": 1.1662929953126032e-05, | |
| "loss": 0.5494, | |
| "step": 200000 | |
| }, | |
| { | |
| "epoch": 23.83, | |
| "learning_rate": 1.1636522083580908e-05, | |
| "loss": 0.5403, | |
| "step": 200500 | |
| }, | |
| { | |
| "epoch": 23.89, | |
| "learning_rate": 1.1610114214035784e-05, | |
| "loss": 0.5522, | |
| "step": 201000 | |
| }, | |
| { | |
| "epoch": 23.95, | |
| "learning_rate": 1.1583759160229748e-05, | |
| "loss": 0.5522, | |
| "step": 201500 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "learning_rate": 1.1557351290684624e-05, | |
| "loss": 0.5405, | |
| "step": 202000 | |
| }, | |
| { | |
| "epoch": 24.06, | |
| "learning_rate": 1.15309434211395e-05, | |
| "loss": 0.5371, | |
| "step": 202500 | |
| }, | |
| { | |
| "epoch": 24.12, | |
| "learning_rate": 1.1504588367333466e-05, | |
| "loss": 0.5378, | |
| "step": 203000 | |
| }, | |
| { | |
| "epoch": 24.18, | |
| "learning_rate": 1.1478180497788342e-05, | |
| "loss": 0.543, | |
| "step": 203500 | |
| }, | |
| { | |
| "epoch": 24.24, | |
| "learning_rate": 1.1451772628243218e-05, | |
| "loss": 0.5359, | |
| "step": 204000 | |
| }, | |
| { | |
| "epoch": 24.3, | |
| "learning_rate": 1.1425364758698093e-05, | |
| "loss": 0.5344, | |
| "step": 204500 | |
| }, | |
| { | |
| "epoch": 24.36, | |
| "learning_rate": 1.1398956889152968e-05, | |
| "loss": 0.548, | |
| "step": 205000 | |
| }, | |
| { | |
| "epoch": 24.42, | |
| "learning_rate": 1.1372549019607844e-05, | |
| "loss": 0.5406, | |
| "step": 205500 | |
| }, | |
| { | |
| "epoch": 24.48, | |
| "learning_rate": 1.134619396580181e-05, | |
| "loss": 0.5398, | |
| "step": 206000 | |
| }, | |
| { | |
| "epoch": 24.54, | |
| "learning_rate": 1.1319786096256687e-05, | |
| "loss": 0.5322, | |
| "step": 206500 | |
| }, | |
| { | |
| "epoch": 24.6, | |
| "learning_rate": 1.1293378226711559e-05, | |
| "loss": 0.5327, | |
| "step": 207000 | |
| }, | |
| { | |
| "epoch": 24.66, | |
| "learning_rate": 1.1266970357166435e-05, | |
| "loss": 0.5395, | |
| "step": 207500 | |
| }, | |
| { | |
| "epoch": 24.72, | |
| "learning_rate": 1.1240615303360403e-05, | |
| "loss": 0.5401, | |
| "step": 208000 | |
| }, | |
| { | |
| "epoch": 24.78, | |
| "learning_rate": 1.1214207433815279e-05, | |
| "loss": 0.5446, | |
| "step": 208500 | |
| }, | |
| { | |
| "epoch": 24.84, | |
| "learning_rate": 1.1187852380009243e-05, | |
| "loss": 0.5431, | |
| "step": 209000 | |
| }, | |
| { | |
| "epoch": 24.9, | |
| "learning_rate": 1.116144451046412e-05, | |
| "loss": 0.543, | |
| "step": 209500 | |
| }, | |
| { | |
| "epoch": 24.96, | |
| "learning_rate": 1.1135036640918995e-05, | |
| "loss": 0.541, | |
| "step": 210000 | |
| }, | |
| { | |
| "epoch": 25.01, | |
| "learning_rate": 1.1108628771373871e-05, | |
| "loss": 0.5417, | |
| "step": 210500 | |
| }, | |
| { | |
| "epoch": 25.07, | |
| "learning_rate": 1.1082273717567836e-05, | |
| "loss": 0.5306, | |
| "step": 211000 | |
| }, | |
| { | |
| "epoch": 25.13, | |
| "learning_rate": 1.1055865848022712e-05, | |
| "loss": 0.5355, | |
| "step": 211500 | |
| }, | |
| { | |
| "epoch": 25.19, | |
| "learning_rate": 1.1029457978477588e-05, | |
| "loss": 0.5358, | |
| "step": 212000 | |
| }, | |
| { | |
| "epoch": 25.25, | |
| "learning_rate": 1.1003050108932463e-05, | |
| "loss": 0.5386, | |
| "step": 212500 | |
| }, | |
| { | |
| "epoch": 25.31, | |
| "learning_rate": 1.0976642239387338e-05, | |
| "loss": 0.54, | |
| "step": 213000 | |
| }, | |
| { | |
| "epoch": 25.37, | |
| "learning_rate": 1.0950287185581304e-05, | |
| "loss": 0.5376, | |
| "step": 213500 | |
| }, | |
| { | |
| "epoch": 25.43, | |
| "learning_rate": 1.0923879316036178e-05, | |
| "loss": 0.5369, | |
| "step": 214000 | |
| }, | |
| { | |
| "epoch": 25.49, | |
| "learning_rate": 1.0897471446491054e-05, | |
| "loss": 0.5352, | |
| "step": 214500 | |
| }, | |
| { | |
| "epoch": 25.55, | |
| "learning_rate": 1.087106357694593e-05, | |
| "loss": 0.5446, | |
| "step": 215000 | |
| }, | |
| { | |
| "epoch": 25.61, | |
| "learning_rate": 1.0844655707400806e-05, | |
| "loss": 0.5399, | |
| "step": 215500 | |
| }, | |
| { | |
| "epoch": 25.67, | |
| "learning_rate": 1.0818247837855682e-05, | |
| "loss": 0.5382, | |
| "step": 216000 | |
| }, | |
| { | |
| "epoch": 25.73, | |
| "learning_rate": 1.0791839968310558e-05, | |
| "loss": 0.5416, | |
| "step": 216500 | |
| }, | |
| { | |
| "epoch": 25.79, | |
| "learning_rate": 1.0765484914504522e-05, | |
| "loss": 0.5414, | |
| "step": 217000 | |
| }, | |
| { | |
| "epoch": 25.85, | |
| "learning_rate": 1.0739077044959398e-05, | |
| "loss": 0.5394, | |
| "step": 217500 | |
| }, | |
| { | |
| "epoch": 25.91, | |
| "learning_rate": 1.0712669175414274e-05, | |
| "loss": 0.5451, | |
| "step": 218000 | |
| }, | |
| { | |
| "epoch": 25.97, | |
| "learning_rate": 1.068626130586915e-05, | |
| "loss": 0.5402, | |
| "step": 218500 | |
| }, | |
| { | |
| "epoch": 26.02, | |
| "learning_rate": 1.0659853436324026e-05, | |
| "loss": 0.5345, | |
| "step": 219000 | |
| }, | |
| { | |
| "epoch": 26.08, | |
| "learning_rate": 1.063349838251799e-05, | |
| "loss": 0.5235, | |
| "step": 219500 | |
| }, | |
| { | |
| "epoch": 26.14, | |
| "learning_rate": 1.0607090512972866e-05, | |
| "loss": 0.5328, | |
| "step": 220000 | |
| }, | |
| { | |
| "epoch": 26.2, | |
| "learning_rate": 1.0580682643427742e-05, | |
| "loss": 0.5376, | |
| "step": 220500 | |
| }, | |
| { | |
| "epoch": 26.26, | |
| "learning_rate": 1.0554274773882618e-05, | |
| "loss": 0.5319, | |
| "step": 221000 | |
| }, | |
| { | |
| "epoch": 26.32, | |
| "learning_rate": 1.0527866904337494e-05, | |
| "loss": 0.5324, | |
| "step": 221500 | |
| }, | |
| { | |
| "epoch": 26.38, | |
| "learning_rate": 1.050145903479237e-05, | |
| "loss": 0.5334, | |
| "step": 222000 | |
| }, | |
| { | |
| "epoch": 26.44, | |
| "learning_rate": 1.0475051165247246e-05, | |
| "loss": 0.5338, | |
| "step": 222500 | |
| }, | |
| { | |
| "epoch": 26.5, | |
| "learning_rate": 1.044864329570212e-05, | |
| "loss": 0.5367, | |
| "step": 223000 | |
| }, | |
| { | |
| "epoch": 26.56, | |
| "learning_rate": 1.0422235426156994e-05, | |
| "loss": 0.5337, | |
| "step": 223500 | |
| }, | |
| { | |
| "epoch": 26.62, | |
| "learning_rate": 1.039582755661187e-05, | |
| "loss": 0.5306, | |
| "step": 224000 | |
| }, | |
| { | |
| "epoch": 26.68, | |
| "learning_rate": 1.0369419687066746e-05, | |
| "loss": 0.5417, | |
| "step": 224500 | |
| }, | |
| { | |
| "epoch": 26.74, | |
| "learning_rate": 1.0343011817521622e-05, | |
| "loss": 0.534, | |
| "step": 225000 | |
| }, | |
| { | |
| "epoch": 26.8, | |
| "learning_rate": 1.0316603947976498e-05, | |
| "loss": 0.5341, | |
| "step": 225500 | |
| }, | |
| { | |
| "epoch": 26.86, | |
| "learning_rate": 1.0290196078431374e-05, | |
| "loss": 0.535, | |
| "step": 226000 | |
| }, | |
| { | |
| "epoch": 26.92, | |
| "learning_rate": 1.0263841024625339e-05, | |
| "loss": 0.5414, | |
| "step": 226500 | |
| }, | |
| { | |
| "epoch": 26.98, | |
| "learning_rate": 1.0237433155080215e-05, | |
| "loss": 0.5375, | |
| "step": 227000 | |
| }, | |
| { | |
| "epoch": 27.04, | |
| "learning_rate": 1.021102528553509e-05, | |
| "loss": 0.5268, | |
| "step": 227500 | |
| }, | |
| { | |
| "epoch": 27.09, | |
| "learning_rate": 1.0184617415989966e-05, | |
| "loss": 0.5277, | |
| "step": 228000 | |
| }, | |
| { | |
| "epoch": 27.15, | |
| "learning_rate": 1.0158262362183933e-05, | |
| "loss": 0.5304, | |
| "step": 228500 | |
| }, | |
| { | |
| "epoch": 27.21, | |
| "learning_rate": 1.0131854492638807e-05, | |
| "loss": 0.5291, | |
| "step": 229000 | |
| }, | |
| { | |
| "epoch": 27.27, | |
| "learning_rate": 1.0105446623093683e-05, | |
| "loss": 0.5287, | |
| "step": 229500 | |
| }, | |
| { | |
| "epoch": 27.33, | |
| "learning_rate": 1.0079091569287649e-05, | |
| "loss": 0.5346, | |
| "step": 230000 | |
| }, | |
| { | |
| "epoch": 27.39, | |
| "learning_rate": 1.0052683699742525e-05, | |
| "loss": 0.5357, | |
| "step": 230500 | |
| }, | |
| { | |
| "epoch": 27.45, | |
| "learning_rate": 1.0026275830197401e-05, | |
| "loss": 0.5317, | |
| "step": 231000 | |
| }, | |
| { | |
| "epoch": 27.51, | |
| "learning_rate": 9.999920776391365e-06, | |
| "loss": 0.5187, | |
| "step": 231500 | |
| }, | |
| { | |
| "epoch": 27.57, | |
| "learning_rate": 9.973512906846241e-06, | |
| "loss": 0.5361, | |
| "step": 232000 | |
| }, | |
| { | |
| "epoch": 27.63, | |
| "learning_rate": 9.947105037301117e-06, | |
| "loss": 0.5361, | |
| "step": 232500 | |
| }, | |
| { | |
| "epoch": 27.69, | |
| "learning_rate": 9.920697167755991e-06, | |
| "loss": 0.5292, | |
| "step": 233000 | |
| }, | |
| { | |
| "epoch": 27.75, | |
| "learning_rate": 9.894289298210867e-06, | |
| "loss": 0.5317, | |
| "step": 233500 | |
| }, | |
| { | |
| "epoch": 27.81, | |
| "learning_rate": 9.867881428665743e-06, | |
| "loss": 0.5294, | |
| "step": 234000 | |
| }, | |
| { | |
| "epoch": 27.87, | |
| "learning_rate": 9.84147355912062e-06, | |
| "loss": 0.5348, | |
| "step": 234500 | |
| }, | |
| { | |
| "epoch": 27.93, | |
| "learning_rate": 9.815065689575495e-06, | |
| "loss": 0.5443, | |
| "step": 235000 | |
| }, | |
| { | |
| "epoch": 27.99, | |
| "learning_rate": 9.788710635769461e-06, | |
| "loss": 0.5355, | |
| "step": 235500 | |
| }, | |
| { | |
| "epoch": 28.05, | |
| "learning_rate": 9.762302766224336e-06, | |
| "loss": 0.5297, | |
| "step": 236000 | |
| }, | |
| { | |
| "epoch": 28.1, | |
| "learning_rate": 9.735894896679212e-06, | |
| "loss": 0.5306, | |
| "step": 236500 | |
| }, | |
| { | |
| "epoch": 28.16, | |
| "learning_rate": 9.709487027134087e-06, | |
| "loss": 0.5251, | |
| "step": 237000 | |
| }, | |
| { | |
| "epoch": 28.22, | |
| "learning_rate": 9.683079157588963e-06, | |
| "loss": 0.5255, | |
| "step": 237500 | |
| }, | |
| { | |
| "epoch": 28.28, | |
| "learning_rate": 9.656671288043838e-06, | |
| "loss": 0.5222, | |
| "step": 238000 | |
| }, | |
| { | |
| "epoch": 28.34, | |
| "learning_rate": 9.630263418498714e-06, | |
| "loss": 0.5281, | |
| "step": 238500 | |
| }, | |
| { | |
| "epoch": 28.4, | |
| "learning_rate": 9.60385554895359e-06, | |
| "loss": 0.5262, | |
| "step": 239000 | |
| }, | |
| { | |
| "epoch": 28.46, | |
| "learning_rate": 9.577447679408465e-06, | |
| "loss": 0.5245, | |
| "step": 239500 | |
| }, | |
| { | |
| "epoch": 28.52, | |
| "learning_rate": 9.551039809863341e-06, | |
| "loss": 0.5321, | |
| "step": 240000 | |
| }, | |
| { | |
| "epoch": 28.58, | |
| "learning_rate": 9.524684756057306e-06, | |
| "loss": 0.5354, | |
| "step": 240500 | |
| }, | |
| { | |
| "epoch": 28.64, | |
| "learning_rate": 9.498276886512182e-06, | |
| "loss": 0.5264, | |
| "step": 241000 | |
| }, | |
| { | |
| "epoch": 28.7, | |
| "learning_rate": 9.471869016967058e-06, | |
| "loss": 0.5244, | |
| "step": 241500 | |
| }, | |
| { | |
| "epoch": 28.76, | |
| "learning_rate": 9.445461147421932e-06, | |
| "loss": 0.5337, | |
| "step": 242000 | |
| }, | |
| { | |
| "epoch": 28.82, | |
| "learning_rate": 9.419106093615898e-06, | |
| "loss": 0.529, | |
| "step": 242500 | |
| }, | |
| { | |
| "epoch": 28.88, | |
| "learning_rate": 9.392698224070774e-06, | |
| "loss": 0.5317, | |
| "step": 243000 | |
| }, | |
| { | |
| "epoch": 28.94, | |
| "learning_rate": 9.366290354525648e-06, | |
| "loss": 0.5285, | |
| "step": 243500 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "learning_rate": 9.339882484980524e-06, | |
| "loss": 0.5338, | |
| "step": 244000 | |
| }, | |
| { | |
| "epoch": 29.06, | |
| "learning_rate": 9.3134746154354e-06, | |
| "loss": 0.5181, | |
| "step": 244500 | |
| }, | |
| { | |
| "epoch": 29.11, | |
| "learning_rate": 9.287066745890276e-06, | |
| "loss": 0.528, | |
| "step": 245000 | |
| }, | |
| { | |
| "epoch": 29.17, | |
| "learning_rate": 9.260658876345152e-06, | |
| "loss": 0.5173, | |
| "step": 245500 | |
| }, | |
| { | |
| "epoch": 29.23, | |
| "learning_rate": 9.234251006800026e-06, | |
| "loss": 0.5261, | |
| "step": 246000 | |
| }, | |
| { | |
| "epoch": 29.29, | |
| "learning_rate": 9.207895952993992e-06, | |
| "loss": 0.5258, | |
| "step": 246500 | |
| }, | |
| { | |
| "epoch": 29.35, | |
| "learning_rate": 9.181488083448868e-06, | |
| "loss": 0.5252, | |
| "step": 247000 | |
| }, | |
| { | |
| "epoch": 29.41, | |
| "learning_rate": 9.155080213903744e-06, | |
| "loss": 0.5298, | |
| "step": 247500 | |
| }, | |
| { | |
| "epoch": 29.47, | |
| "learning_rate": 9.12867234435862e-06, | |
| "loss": 0.519, | |
| "step": 248000 | |
| }, | |
| { | |
| "epoch": 29.53, | |
| "learning_rate": 9.102317290552586e-06, | |
| "loss": 0.5197, | |
| "step": 248500 | |
| }, | |
| { | |
| "epoch": 29.59, | |
| "learning_rate": 9.07590942100746e-06, | |
| "loss": 0.5199, | |
| "step": 249000 | |
| }, | |
| { | |
| "epoch": 29.65, | |
| "learning_rate": 9.049501551462337e-06, | |
| "loss": 0.5279, | |
| "step": 249500 | |
| }, | |
| { | |
| "epoch": 29.71, | |
| "learning_rate": 9.023093681917212e-06, | |
| "loss": 0.5245, | |
| "step": 250000 | |
| }, | |
| { | |
| "epoch": 29.77, | |
| "learning_rate": 8.996738628111179e-06, | |
| "loss": 0.5243, | |
| "step": 250500 | |
| }, | |
| { | |
| "epoch": 29.83, | |
| "learning_rate": 8.970330758566053e-06, | |
| "loss": 0.5296, | |
| "step": 251000 | |
| }, | |
| { | |
| "epoch": 29.89, | |
| "learning_rate": 8.94397570476002e-06, | |
| "loss": 0.5266, | |
| "step": 251500 | |
| }, | |
| { | |
| "epoch": 29.95, | |
| "learning_rate": 8.917567835214893e-06, | |
| "loss": 0.5282, | |
| "step": 252000 | |
| }, | |
| { | |
| "epoch": 30.01, | |
| "learning_rate": 8.89115996566977e-06, | |
| "loss": 0.5299, | |
| "step": 252500 | |
| }, | |
| { | |
| "epoch": 30.07, | |
| "learning_rate": 8.864752096124645e-06, | |
| "loss": 0.5238, | |
| "step": 253000 | |
| }, | |
| { | |
| "epoch": 30.12, | |
| "learning_rate": 8.838344226579521e-06, | |
| "loss": 0.5256, | |
| "step": 253500 | |
| }, | |
| { | |
| "epoch": 30.18, | |
| "learning_rate": 8.811936357034397e-06, | |
| "loss": 0.522, | |
| "step": 254000 | |
| }, | |
| { | |
| "epoch": 30.24, | |
| "learning_rate": 8.785528487489271e-06, | |
| "loss": 0.5249, | |
| "step": 254500 | |
| }, | |
| { | |
| "epoch": 30.3, | |
| "learning_rate": 8.759120617944147e-06, | |
| "loss": 0.5206, | |
| "step": 255000 | |
| }, | |
| { | |
| "epoch": 30.36, | |
| "learning_rate": 8.732712748399023e-06, | |
| "loss": 0.5193, | |
| "step": 255500 | |
| }, | |
| { | |
| "epoch": 30.42, | |
| "learning_rate": 8.70635769459299e-06, | |
| "loss": 0.529, | |
| "step": 256000 | |
| }, | |
| { | |
| "epoch": 30.48, | |
| "learning_rate": 8.680002640786956e-06, | |
| "loss": 0.5159, | |
| "step": 256500 | |
| }, | |
| { | |
| "epoch": 30.54, | |
| "learning_rate": 8.653594771241832e-06, | |
| "loss": 0.5337, | |
| "step": 257000 | |
| }, | |
| { | |
| "epoch": 30.6, | |
| "learning_rate": 8.627186901696706e-06, | |
| "loss": 0.5298, | |
| "step": 257500 | |
| }, | |
| { | |
| "epoch": 30.66, | |
| "learning_rate": 8.600779032151582e-06, | |
| "loss": 0.5296, | |
| "step": 258000 | |
| }, | |
| { | |
| "epoch": 30.72, | |
| "learning_rate": 8.574371162606458e-06, | |
| "loss": 0.5227, | |
| "step": 258500 | |
| }, | |
| { | |
| "epoch": 30.78, | |
| "learning_rate": 8.547963293061334e-06, | |
| "loss": 0.5296, | |
| "step": 259000 | |
| }, | |
| { | |
| "epoch": 30.84, | |
| "learning_rate": 8.52155542351621e-06, | |
| "loss": 0.53, | |
| "step": 259500 | |
| }, | |
| { | |
| "epoch": 30.9, | |
| "learning_rate": 8.495147553971084e-06, | |
| "loss": 0.5256, | |
| "step": 260000 | |
| }, | |
| { | |
| "epoch": 30.96, | |
| "learning_rate": 8.46873968442596e-06, | |
| "loss": 0.5272, | |
| "step": 260500 | |
| }, | |
| { | |
| "epoch": 31.02, | |
| "learning_rate": 8.442384630619926e-06, | |
| "loss": 0.5294, | |
| "step": 261000 | |
| }, | |
| { | |
| "epoch": 31.08, | |
| "learning_rate": 8.415976761074802e-06, | |
| "loss": 0.511, | |
| "step": 261500 | |
| }, | |
| { | |
| "epoch": 31.13, | |
| "learning_rate": 8.389568891529678e-06, | |
| "loss": 0.5233, | |
| "step": 262000 | |
| }, | |
| { | |
| "epoch": 31.19, | |
| "learning_rate": 8.363161021984552e-06, | |
| "loss": 0.5217, | |
| "step": 262500 | |
| }, | |
| { | |
| "epoch": 31.25, | |
| "learning_rate": 8.336753152439428e-06, | |
| "loss": 0.5162, | |
| "step": 263000 | |
| }, | |
| { | |
| "epoch": 31.31, | |
| "learning_rate": 8.310345282894304e-06, | |
| "loss": 0.5129, | |
| "step": 263500 | |
| }, | |
| { | |
| "epoch": 31.37, | |
| "learning_rate": 8.28393741334918e-06, | |
| "loss": 0.5178, | |
| "step": 264000 | |
| }, | |
| { | |
| "epoch": 31.43, | |
| "learning_rate": 8.257529543804056e-06, | |
| "loss": 0.5155, | |
| "step": 264500 | |
| }, | |
| { | |
| "epoch": 31.49, | |
| "learning_rate": 8.23112167425893e-06, | |
| "loss": 0.5211, | |
| "step": 265000 | |
| }, | |
| { | |
| "epoch": 31.55, | |
| "learning_rate": 8.204819436191986e-06, | |
| "loss": 0.5248, | |
| "step": 265500 | |
| }, | |
| { | |
| "epoch": 31.61, | |
| "learning_rate": 8.17841156664686e-06, | |
| "loss": 0.5344, | |
| "step": 266000 | |
| }, | |
| { | |
| "epoch": 31.67, | |
| "learning_rate": 8.152003697101737e-06, | |
| "loss": 0.5237, | |
| "step": 266500 | |
| }, | |
| { | |
| "epoch": 31.73, | |
| "learning_rate": 8.125595827556612e-06, | |
| "loss": 0.5172, | |
| "step": 267000 | |
| }, | |
| { | |
| "epoch": 31.79, | |
| "learning_rate": 8.099240773750579e-06, | |
| "loss": 0.5237, | |
| "step": 267500 | |
| }, | |
| { | |
| "epoch": 31.85, | |
| "learning_rate": 8.072832904205455e-06, | |
| "loss": 0.5229, | |
| "step": 268000 | |
| }, | |
| { | |
| "epoch": 31.91, | |
| "learning_rate": 8.046425034660329e-06, | |
| "loss": 0.5186, | |
| "step": 268500 | |
| }, | |
| { | |
| "epoch": 31.97, | |
| "learning_rate": 8.020069980854295e-06, | |
| "loss": 0.5287, | |
| "step": 269000 | |
| }, | |
| { | |
| "epoch": 32.03, | |
| "learning_rate": 7.993662111309171e-06, | |
| "loss": 0.527, | |
| "step": 269500 | |
| }, | |
| { | |
| "epoch": 32.09, | |
| "learning_rate": 7.967254241764047e-06, | |
| "loss": 0.5139, | |
| "step": 270000 | |
| }, | |
| { | |
| "epoch": 32.14, | |
| "learning_rate": 7.940846372218923e-06, | |
| "loss": 0.5184, | |
| "step": 270500 | |
| }, | |
| { | |
| "epoch": 32.2, | |
| "learning_rate": 7.914438502673799e-06, | |
| "loss": 0.5189, | |
| "step": 271000 | |
| }, | |
| { | |
| "epoch": 32.26, | |
| "learning_rate": 7.888030633128673e-06, | |
| "loss": 0.5221, | |
| "step": 271500 | |
| }, | |
| { | |
| "epoch": 32.32, | |
| "learning_rate": 7.861622763583549e-06, | |
| "loss": 0.5176, | |
| "step": 272000 | |
| }, | |
| { | |
| "epoch": 32.38, | |
| "learning_rate": 7.835267709777515e-06, | |
| "loss": 0.5134, | |
| "step": 272500 | |
| }, | |
| { | |
| "epoch": 32.44, | |
| "learning_rate": 7.80885984023239e-06, | |
| "loss": 0.5187, | |
| "step": 273000 | |
| }, | |
| { | |
| "epoch": 32.5, | |
| "learning_rate": 7.782451970687265e-06, | |
| "loss": 0.522, | |
| "step": 273500 | |
| }, | |
| { | |
| "epoch": 32.56, | |
| "learning_rate": 7.756044101142141e-06, | |
| "loss": 0.5151, | |
| "step": 274000 | |
| }, | |
| { | |
| "epoch": 32.62, | |
| "learning_rate": 7.729636231597017e-06, | |
| "loss": 0.5237, | |
| "step": 274500 | |
| }, | |
| { | |
| "epoch": 32.68, | |
| "learning_rate": 7.703228362051893e-06, | |
| "loss": 0.5176, | |
| "step": 275000 | |
| }, | |
| { | |
| "epoch": 32.74, | |
| "learning_rate": 7.676820492506767e-06, | |
| "loss": 0.5227, | |
| "step": 275500 | |
| }, | |
| { | |
| "epoch": 32.8, | |
| "learning_rate": 7.650412622961643e-06, | |
| "loss": 0.5162, | |
| "step": 276000 | |
| }, | |
| { | |
| "epoch": 32.86, | |
| "learning_rate": 7.624004753416518e-06, | |
| "loss": 0.521, | |
| "step": 276500 | |
| }, | |
| { | |
| "epoch": 32.92, | |
| "learning_rate": 7.597596883871394e-06, | |
| "loss": 0.5231, | |
| "step": 277000 | |
| }, | |
| { | |
| "epoch": 32.98, | |
| "learning_rate": 7.57124183006536e-06, | |
| "loss": 0.5198, | |
| "step": 277500 | |
| }, | |
| { | |
| "epoch": 33.04, | |
| "learning_rate": 7.5448339605202355e-06, | |
| "loss": 0.5086, | |
| "step": 278000 | |
| }, | |
| { | |
| "epoch": 33.1, | |
| "learning_rate": 7.518426090975111e-06, | |
| "loss": 0.5124, | |
| "step": 278500 | |
| }, | |
| { | |
| "epoch": 33.16, | |
| "learning_rate": 7.4920182214299865e-06, | |
| "loss": 0.5175, | |
| "step": 279000 | |
| }, | |
| { | |
| "epoch": 33.21, | |
| "learning_rate": 7.465663167623952e-06, | |
| "loss": 0.5131, | |
| "step": 279500 | |
| }, | |
| { | |
| "epoch": 33.27, | |
| "learning_rate": 7.439255298078828e-06, | |
| "loss": 0.5114, | |
| "step": 280000 | |
| }, | |
| { | |
| "epoch": 33.33, | |
| "learning_rate": 7.412847428533704e-06, | |
| "loss": 0.5118, | |
| "step": 280500 | |
| }, | |
| { | |
| "epoch": 33.39, | |
| "learning_rate": 7.38643955898858e-06, | |
| "loss": 0.5116, | |
| "step": 281000 | |
| }, | |
| { | |
| "epoch": 33.45, | |
| "learning_rate": 7.360084505182545e-06, | |
| "loss": 0.5197, | |
| "step": 281500 | |
| }, | |
| { | |
| "epoch": 33.51, | |
| "learning_rate": 7.33367663563742e-06, | |
| "loss": 0.5161, | |
| "step": 282000 | |
| }, | |
| { | |
| "epoch": 33.57, | |
| "learning_rate": 7.307268766092296e-06, | |
| "loss": 0.5166, | |
| "step": 282500 | |
| }, | |
| { | |
| "epoch": 33.63, | |
| "learning_rate": 7.280913712286261e-06, | |
| "loss": 0.5203, | |
| "step": 283000 | |
| }, | |
| { | |
| "epoch": 33.69, | |
| "learning_rate": 7.254505842741137e-06, | |
| "loss": 0.513, | |
| "step": 283500 | |
| }, | |
| { | |
| "epoch": 33.75, | |
| "learning_rate": 7.228097973196013e-06, | |
| "loss": 0.5139, | |
| "step": 284000 | |
| }, | |
| { | |
| "epoch": 33.81, | |
| "learning_rate": 7.201690103650889e-06, | |
| "loss": 0.5123, | |
| "step": 284500 | |
| }, | |
| { | |
| "epoch": 33.87, | |
| "learning_rate": 7.175282234105763e-06, | |
| "loss": 0.5191, | |
| "step": 285000 | |
| }, | |
| { | |
| "epoch": 33.93, | |
| "learning_rate": 7.148874364560639e-06, | |
| "loss": 0.5228, | |
| "step": 285500 | |
| }, | |
| { | |
| "epoch": 33.99, | |
| "learning_rate": 7.122466495015515e-06, | |
| "loss": 0.5165, | |
| "step": 286000 | |
| }, | |
| { | |
| "epoch": 34.05, | |
| "learning_rate": 7.096058625470391e-06, | |
| "loss": 0.5089, | |
| "step": 286500 | |
| }, | |
| { | |
| "epoch": 34.11, | |
| "learning_rate": 7.069650755925266e-06, | |
| "loss": 0.5066, | |
| "step": 287000 | |
| }, | |
| { | |
| "epoch": 34.17, | |
| "learning_rate": 7.043242886380141e-06, | |
| "loss": 0.514, | |
| "step": 287500 | |
| }, | |
| { | |
| "epoch": 34.22, | |
| "learning_rate": 7.0168878325741075e-06, | |
| "loss": 0.5099, | |
| "step": 288000 | |
| }, | |
| { | |
| "epoch": 34.28, | |
| "learning_rate": 6.9904799630289834e-06, | |
| "loss": 0.5124, | |
| "step": 288500 | |
| }, | |
| { | |
| "epoch": 34.34, | |
| "learning_rate": 6.964072093483859e-06, | |
| "loss": 0.5097, | |
| "step": 289000 | |
| }, | |
| { | |
| "epoch": 34.4, | |
| "learning_rate": 6.9376642239387344e-06, | |
| "loss": 0.5155, | |
| "step": 289500 | |
| }, | |
| { | |
| "epoch": 34.46, | |
| "learning_rate": 6.9112563543936095e-06, | |
| "loss": 0.5146, | |
| "step": 290000 | |
| }, | |
| { | |
| "epoch": 34.52, | |
| "learning_rate": 6.8848484848484854e-06, | |
| "loss": 0.5178, | |
| "step": 290500 | |
| }, | |
| { | |
| "epoch": 34.58, | |
| "learning_rate": 6.858493431042451e-06, | |
| "loss": 0.5158, | |
| "step": 291000 | |
| }, | |
| { | |
| "epoch": 34.64, | |
| "learning_rate": 6.832085561497327e-06, | |
| "loss": 0.5156, | |
| "step": 291500 | |
| }, | |
| { | |
| "epoch": 34.7, | |
| "learning_rate": 6.805677691952203e-06, | |
| "loss": 0.5132, | |
| "step": 292000 | |
| }, | |
| { | |
| "epoch": 34.76, | |
| "learning_rate": 6.779322638146169e-06, | |
| "loss": 0.5171, | |
| "step": 292500 | |
| }, | |
| { | |
| "epoch": 34.82, | |
| "learning_rate": 6.752914768601043e-06, | |
| "loss": 0.5186, | |
| "step": 293000 | |
| }, | |
| { | |
| "epoch": 34.88, | |
| "learning_rate": 6.726506899055919e-06, | |
| "loss": 0.5098, | |
| "step": 293500 | |
| }, | |
| { | |
| "epoch": 34.94, | |
| "learning_rate": 6.700099029510795e-06, | |
| "loss": 0.5044, | |
| "step": 294000 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "learning_rate": 6.673691159965671e-06, | |
| "loss": 0.5106, | |
| "step": 294500 | |
| }, | |
| { | |
| "epoch": 35.06, | |
| "learning_rate": 6.647336106159636e-06, | |
| "loss": 0.5083, | |
| "step": 295000 | |
| }, | |
| { | |
| "epoch": 35.12, | |
| "learning_rate": 6.620928236614511e-06, | |
| "loss": 0.5165, | |
| "step": 295500 | |
| }, | |
| { | |
| "epoch": 35.18, | |
| "learning_rate": 6.594520367069387e-06, | |
| "loss": 0.5091, | |
| "step": 296000 | |
| }, | |
| { | |
| "epoch": 35.23, | |
| "learning_rate": 6.568112497524262e-06, | |
| "loss": 0.506, | |
| "step": 296500 | |
| }, | |
| { | |
| "epoch": 35.29, | |
| "learning_rate": 6.541704627979138e-06, | |
| "loss": 0.5113, | |
| "step": 297000 | |
| }, | |
| { | |
| "epoch": 35.35, | |
| "learning_rate": 6.515296758434014e-06, | |
| "loss": 0.5108, | |
| "step": 297500 | |
| }, | |
| { | |
| "epoch": 35.41, | |
| "learning_rate": 6.488888888888889e-06, | |
| "loss": 0.5115, | |
| "step": 298000 | |
| }, | |
| { | |
| "epoch": 35.47, | |
| "learning_rate": 6.462533835082855e-06, | |
| "loss": 0.512, | |
| "step": 298500 | |
| }, | |
| { | |
| "epoch": 35.53, | |
| "learning_rate": 6.4361259655377306e-06, | |
| "loss": 0.5078, | |
| "step": 299000 | |
| }, | |
| { | |
| "epoch": 35.59, | |
| "learning_rate": 6.4097180959926065e-06, | |
| "loss": 0.5146, | |
| "step": 299500 | |
| }, | |
| { | |
| "epoch": 35.65, | |
| "learning_rate": 6.383310226447482e-06, | |
| "loss": 0.5133, | |
| "step": 300000 | |
| }, | |
| { | |
| "epoch": 35.71, | |
| "learning_rate": 6.356902356902357e-06, | |
| "loss": 0.5201, | |
| "step": 300500 | |
| }, | |
| { | |
| "epoch": 35.77, | |
| "learning_rate": 6.3304944873572326e-06, | |
| "loss": 0.5132, | |
| "step": 301000 | |
| }, | |
| { | |
| "epoch": 35.83, | |
| "learning_rate": 6.304139433551199e-06, | |
| "loss": 0.5139, | |
| "step": 301500 | |
| }, | |
| { | |
| "epoch": 35.89, | |
| "learning_rate": 6.277731564006075e-06, | |
| "loss": 0.5084, | |
| "step": 302000 | |
| }, | |
| { | |
| "epoch": 35.95, | |
| "learning_rate": 6.25132369446095e-06, | |
| "loss": 0.5155, | |
| "step": 302500 | |
| }, | |
| { | |
| "epoch": 36.01, | |
| "learning_rate": 6.224915824915826e-06, | |
| "loss": 0.5168, | |
| "step": 303000 | |
| }, | |
| { | |
| "epoch": 36.07, | |
| "learning_rate": 6.198507955370701e-06, | |
| "loss": 0.5004, | |
| "step": 303500 | |
| }, | |
| { | |
| "epoch": 36.13, | |
| "learning_rate": 6.172100085825577e-06, | |
| "loss": 0.5142, | |
| "step": 304000 | |
| }, | |
| { | |
| "epoch": 36.19, | |
| "learning_rate": 6.1458506634977235e-06, | |
| "loss": 0.5038, | |
| "step": 304500 | |
| }, | |
| { | |
| "epoch": 36.24, | |
| "learning_rate": 6.119442793952598e-06, | |
| "loss": 0.5157, | |
| "step": 305000 | |
| }, | |
| { | |
| "epoch": 36.3, | |
| "learning_rate": 6.093034924407474e-06, | |
| "loss": 0.5114, | |
| "step": 305500 | |
| }, | |
| { | |
| "epoch": 36.36, | |
| "learning_rate": 6.06662705486235e-06, | |
| "loss": 0.5136, | |
| "step": 306000 | |
| }, | |
| { | |
| "epoch": 36.42, | |
| "learning_rate": 6.0402191853172255e-06, | |
| "loss": 0.5034, | |
| "step": 306500 | |
| }, | |
| { | |
| "epoch": 36.48, | |
| "learning_rate": 6.0138113157721e-06, | |
| "loss": 0.517, | |
| "step": 307000 | |
| }, | |
| { | |
| "epoch": 36.54, | |
| "learning_rate": 5.987456261966066e-06, | |
| "loss": 0.516, | |
| "step": 307500 | |
| }, | |
| { | |
| "epoch": 36.6, | |
| "learning_rate": 5.961048392420942e-06, | |
| "loss": 0.5104, | |
| "step": 308000 | |
| }, | |
| { | |
| "epoch": 36.66, | |
| "learning_rate": 5.934640522875818e-06, | |
| "loss": 0.5108, | |
| "step": 308500 | |
| }, | |
| { | |
| "epoch": 36.72, | |
| "learning_rate": 5.908232653330693e-06, | |
| "loss": 0.5094, | |
| "step": 309000 | |
| }, | |
| { | |
| "epoch": 36.78, | |
| "learning_rate": 5.881824783785569e-06, | |
| "loss": 0.5145, | |
| "step": 309500 | |
| }, | |
| { | |
| "epoch": 36.84, | |
| "learning_rate": 5.855416914240444e-06, | |
| "loss": 0.5176, | |
| "step": 310000 | |
| }, | |
| { | |
| "epoch": 36.9, | |
| "learning_rate": 5.82900904469532e-06, | |
| "loss": 0.5082, | |
| "step": 310500 | |
| }, | |
| { | |
| "epoch": 36.96, | |
| "learning_rate": 5.802601175150196e-06, | |
| "loss": 0.5141, | |
| "step": 311000 | |
| }, | |
| { | |
| "epoch": 37.02, | |
| "learning_rate": 5.776193305605071e-06, | |
| "loss": 0.5116, | |
| "step": 311500 | |
| }, | |
| { | |
| "epoch": 37.08, | |
| "learning_rate": 5.749785436059946e-06, | |
| "loss": 0.5058, | |
| "step": 312000 | |
| }, | |
| { | |
| "epoch": 37.14, | |
| "learning_rate": 5.723377566514822e-06, | |
| "loss": 0.5042, | |
| "step": 312500 | |
| }, | |
| { | |
| "epoch": 37.2, | |
| "learning_rate": 5.696969696969698e-06, | |
| "loss": 0.5097, | |
| "step": 313000 | |
| }, | |
| { | |
| "epoch": 37.25, | |
| "learning_rate": 5.670614643163663e-06, | |
| "loss": 0.5118, | |
| "step": 313500 | |
| }, | |
| { | |
| "epoch": 37.31, | |
| "learning_rate": 5.644206773618539e-06, | |
| "loss": 0.5056, | |
| "step": 314000 | |
| }, | |
| { | |
| "epoch": 37.37, | |
| "learning_rate": 5.617798904073415e-06, | |
| "loss": 0.5143, | |
| "step": 314500 | |
| }, | |
| { | |
| "epoch": 37.43, | |
| "learning_rate": 5.59144385026738e-06, | |
| "loss": 0.516, | |
| "step": 315000 | |
| }, | |
| { | |
| "epoch": 37.49, | |
| "learning_rate": 5.565035980722255e-06, | |
| "loss": 0.5028, | |
| "step": 315500 | |
| }, | |
| { | |
| "epoch": 37.55, | |
| "learning_rate": 5.538628111177131e-06, | |
| "loss": 0.5079, | |
| "step": 316000 | |
| }, | |
| { | |
| "epoch": 37.61, | |
| "learning_rate": 5.512220241632007e-06, | |
| "loss": 0.5101, | |
| "step": 316500 | |
| }, | |
| { | |
| "epoch": 37.67, | |
| "learning_rate": 5.485812372086883e-06, | |
| "loss": 0.5101, | |
| "step": 317000 | |
| }, | |
| { | |
| "epoch": 37.73, | |
| "learning_rate": 5.459404502541757e-06, | |
| "loss": 0.5077, | |
| "step": 317500 | |
| }, | |
| { | |
| "epoch": 37.79, | |
| "learning_rate": 5.432996632996633e-06, | |
| "loss": 0.5094, | |
| "step": 318000 | |
| }, | |
| { | |
| "epoch": 37.85, | |
| "learning_rate": 5.406588763451509e-06, | |
| "loss": 0.5128, | |
| "step": 318500 | |
| }, | |
| { | |
| "epoch": 37.91, | |
| "learning_rate": 5.380180893906385e-06, | |
| "loss": 0.5121, | |
| "step": 319000 | |
| }, | |
| { | |
| "epoch": 37.97, | |
| "learning_rate": 5.353773024361259e-06, | |
| "loss": 0.5071, | |
| "step": 319500 | |
| }, | |
| { | |
| "epoch": 38.03, | |
| "learning_rate": 5.327365154816135e-06, | |
| "loss": 0.5088, | |
| "step": 320000 | |
| }, | |
| { | |
| "epoch": 38.09, | |
| "learning_rate": 5.300957285271011e-06, | |
| "loss": 0.5037, | |
| "step": 320500 | |
| }, | |
| { | |
| "epoch": 38.15, | |
| "learning_rate": 5.274549415725887e-06, | |
| "loss": 0.5083, | |
| "step": 321000 | |
| }, | |
| { | |
| "epoch": 38.21, | |
| "learning_rate": 5.248141546180763e-06, | |
| "loss": 0.5101, | |
| "step": 321500 | |
| }, | |
| { | |
| "epoch": 38.26, | |
| "learning_rate": 5.221733676635637e-06, | |
| "loss": 0.5046, | |
| "step": 322000 | |
| }, | |
| { | |
| "epoch": 38.32, | |
| "learning_rate": 5.195325807090513e-06, | |
| "loss": 0.5021, | |
| "step": 322500 | |
| }, | |
| { | |
| "epoch": 38.38, | |
| "learning_rate": 5.169023569023569e-06, | |
| "loss": 0.5116, | |
| "step": 323000 | |
| }, | |
| { | |
| "epoch": 38.44, | |
| "learning_rate": 5.142615699478445e-06, | |
| "loss": 0.5018, | |
| "step": 323500 | |
| }, | |
| { | |
| "epoch": 38.5, | |
| "learning_rate": 5.116207829933321e-06, | |
| "loss": 0.5087, | |
| "step": 324000 | |
| }, | |
| { | |
| "epoch": 38.56, | |
| "learning_rate": 5.089799960388197e-06, | |
| "loss": 0.5033, | |
| "step": 324500 | |
| }, | |
| { | |
| "epoch": 38.62, | |
| "learning_rate": 5.063392090843071e-06, | |
| "loss": 0.515, | |
| "step": 325000 | |
| }, | |
| { | |
| "epoch": 38.68, | |
| "learning_rate": 5.036984221297947e-06, | |
| "loss": 0.5031, | |
| "step": 325500 | |
| }, | |
| { | |
| "epoch": 38.74, | |
| "learning_rate": 5.010576351752823e-06, | |
| "loss": 0.5077, | |
| "step": 326000 | |
| }, | |
| { | |
| "epoch": 38.8, | |
| "learning_rate": 4.984168482207699e-06, | |
| "loss": 0.5142, | |
| "step": 326500 | |
| }, | |
| { | |
| "epoch": 38.86, | |
| "learning_rate": 4.957760612662574e-06, | |
| "loss": 0.5064, | |
| "step": 327000 | |
| }, | |
| { | |
| "epoch": 38.92, | |
| "learning_rate": 4.93135274311745e-06, | |
| "loss": 0.5103, | |
| "step": 327500 | |
| }, | |
| { | |
| "epoch": 38.98, | |
| "learning_rate": 4.904944873572325e-06, | |
| "loss": 0.5062, | |
| "step": 328000 | |
| }, | |
| { | |
| "epoch": 39.04, | |
| "learning_rate": 4.878537004027201e-06, | |
| "loss": 0.5077, | |
| "step": 328500 | |
| }, | |
| { | |
| "epoch": 39.1, | |
| "learning_rate": 4.852181950221166e-06, | |
| "loss": 0.4985, | |
| "step": 329000 | |
| }, | |
| { | |
| "epoch": 39.16, | |
| "learning_rate": 4.825774080676042e-06, | |
| "loss": 0.4965, | |
| "step": 329500 | |
| }, | |
| { | |
| "epoch": 39.22, | |
| "learning_rate": 4.799419026870007e-06, | |
| "loss": 0.5059, | |
| "step": 330000 | |
| }, | |
| { | |
| "epoch": 39.28, | |
| "learning_rate": 4.773011157324883e-06, | |
| "loss": 0.5083, | |
| "step": 330500 | |
| }, | |
| { | |
| "epoch": 39.33, | |
| "learning_rate": 4.746603287779758e-06, | |
| "loss": 0.4982, | |
| "step": 331000 | |
| }, | |
| { | |
| "epoch": 39.39, | |
| "learning_rate": 4.720195418234634e-06, | |
| "loss": 0.5043, | |
| "step": 331500 | |
| }, | |
| { | |
| "epoch": 39.45, | |
| "learning_rate": 4.6938403644286005e-06, | |
| "loss": 0.5067, | |
| "step": 332000 | |
| }, | |
| { | |
| "epoch": 39.51, | |
| "learning_rate": 4.667432494883476e-06, | |
| "loss": 0.5068, | |
| "step": 332500 | |
| }, | |
| { | |
| "epoch": 39.57, | |
| "learning_rate": 4.6410246253383515e-06, | |
| "loss": 0.5096, | |
| "step": 333000 | |
| }, | |
| { | |
| "epoch": 39.63, | |
| "learning_rate": 4.614616755793227e-06, | |
| "loss": 0.5103, | |
| "step": 333500 | |
| }, | |
| { | |
| "epoch": 39.69, | |
| "learning_rate": 4.5882088862481025e-06, | |
| "loss": 0.5069, | |
| "step": 334000 | |
| }, | |
| { | |
| "epoch": 39.75, | |
| "learning_rate": 4.561853832442068e-06, | |
| "loss": 0.5107, | |
| "step": 334500 | |
| }, | |
| { | |
| "epoch": 39.81, | |
| "learning_rate": 4.535445962896944e-06, | |
| "loss": 0.5059, | |
| "step": 335000 | |
| }, | |
| { | |
| "epoch": 39.87, | |
| "learning_rate": 4.509038093351819e-06, | |
| "loss": 0.51, | |
| "step": 335500 | |
| }, | |
| { | |
| "epoch": 39.93, | |
| "learning_rate": 4.482630223806695e-06, | |
| "loss": 0.5019, | |
| "step": 336000 | |
| }, | |
| { | |
| "epoch": 39.99, | |
| "learning_rate": 4.45622235426157e-06, | |
| "loss": 0.505, | |
| "step": 336500 | |
| }, | |
| { | |
| "epoch": 40.05, | |
| "learning_rate": 4.429867300455536e-06, | |
| "loss": 0.4982, | |
| "step": 337000 | |
| }, | |
| { | |
| "epoch": 40.11, | |
| "learning_rate": 4.403459430910412e-06, | |
| "loss": 0.5016, | |
| "step": 337500 | |
| }, | |
| { | |
| "epoch": 40.17, | |
| "learning_rate": 4.377051561365287e-06, | |
| "loss": 0.5085, | |
| "step": 338000 | |
| }, | |
| { | |
| "epoch": 40.23, | |
| "learning_rate": 4.350643691820163e-06, | |
| "loss": 0.5048, | |
| "step": 338500 | |
| }, | |
| { | |
| "epoch": 40.29, | |
| "learning_rate": 4.324288638014128e-06, | |
| "loss": 0.4955, | |
| "step": 339000 | |
| }, | |
| { | |
| "epoch": 40.34, | |
| "learning_rate": 4.297880768469004e-06, | |
| "loss": 0.4997, | |
| "step": 339500 | |
| }, | |
| { | |
| "epoch": 40.4, | |
| "learning_rate": 4.271472898923879e-06, | |
| "loss": 0.5124, | |
| "step": 340000 | |
| }, | |
| { | |
| "epoch": 40.46, | |
| "learning_rate": 4.245065029378755e-06, | |
| "loss": 0.511, | |
| "step": 340500 | |
| }, | |
| { | |
| "epoch": 40.52, | |
| "learning_rate": 4.2187099755727216e-06, | |
| "loss": 0.5061, | |
| "step": 341000 | |
| }, | |
| { | |
| "epoch": 40.58, | |
| "learning_rate": 4.192302106027597e-06, | |
| "loss": 0.5064, | |
| "step": 341500 | |
| }, | |
| { | |
| "epoch": 40.64, | |
| "learning_rate": 4.165947052221562e-06, | |
| "loss": 0.5041, | |
| "step": 342000 | |
| }, | |
| { | |
| "epoch": 40.7, | |
| "learning_rate": 4.139539182676438e-06, | |
| "loss": 0.5062, | |
| "step": 342500 | |
| }, | |
| { | |
| "epoch": 40.76, | |
| "learning_rate": 4.113131313131313e-06, | |
| "loss": 0.5009, | |
| "step": 343000 | |
| }, | |
| { | |
| "epoch": 40.82, | |
| "learning_rate": 4.086723443586189e-06, | |
| "loss": 0.511, | |
| "step": 343500 | |
| }, | |
| { | |
| "epoch": 40.88, | |
| "learning_rate": 4.060315574041064e-06, | |
| "loss": 0.5067, | |
| "step": 344000 | |
| }, | |
| { | |
| "epoch": 40.94, | |
| "learning_rate": 4.03396052023503e-06, | |
| "loss": 0.5067, | |
| "step": 344500 | |
| }, | |
| { | |
| "epoch": 41.0, | |
| "learning_rate": 4.007552650689906e-06, | |
| "loss": 0.5056, | |
| "step": 345000 | |
| }, | |
| { | |
| "epoch": 41.06, | |
| "learning_rate": 3.981144781144781e-06, | |
| "loss": 0.4987, | |
| "step": 345500 | |
| }, | |
| { | |
| "epoch": 41.12, | |
| "learning_rate": 3.954736911599657e-06, | |
| "loss": 0.499, | |
| "step": 346000 | |
| }, | |
| { | |
| "epoch": 41.18, | |
| "learning_rate": 3.928329042054533e-06, | |
| "loss": 0.5054, | |
| "step": 346500 | |
| }, | |
| { | |
| "epoch": 41.24, | |
| "learning_rate": 3.901921172509408e-06, | |
| "loss": 0.501, | |
| "step": 347000 | |
| }, | |
| { | |
| "epoch": 41.3, | |
| "learning_rate": 3.875513302964284e-06, | |
| "loss": 0.5058, | |
| "step": 347500 | |
| }, | |
| { | |
| "epoch": 41.35, | |
| "learning_rate": 3.849105433419159e-06, | |
| "loss": 0.4985, | |
| "step": 348000 | |
| }, | |
| { | |
| "epoch": 41.41, | |
| "learning_rate": 3.822697563874035e-06, | |
| "loss": 0.5019, | |
| "step": 348500 | |
| }, | |
| { | |
| "epoch": 41.47, | |
| "learning_rate": 3.7963953258070908e-06, | |
| "loss": 0.508, | |
| "step": 349000 | |
| }, | |
| { | |
| "epoch": 41.53, | |
| "learning_rate": 3.7699874562619667e-06, | |
| "loss": 0.5063, | |
| "step": 349500 | |
| }, | |
| { | |
| "epoch": 41.59, | |
| "learning_rate": 3.7435795867168418e-06, | |
| "loss": 0.5048, | |
| "step": 350000 | |
| }, | |
| { | |
| "epoch": 41.65, | |
| "learning_rate": 3.7171717171717177e-06, | |
| "loss": 0.5026, | |
| "step": 350500 | |
| }, | |
| { | |
| "epoch": 41.71, | |
| "learning_rate": 3.6907638476265928e-06, | |
| "loss": 0.504, | |
| "step": 351000 | |
| }, | |
| { | |
| "epoch": 41.77, | |
| "learning_rate": 3.6643559780814687e-06, | |
| "loss": 0.5011, | |
| "step": 351500 | |
| }, | |
| { | |
| "epoch": 41.83, | |
| "learning_rate": 3.637948108536344e-06, | |
| "loss": 0.503, | |
| "step": 352000 | |
| }, | |
| { | |
| "epoch": 41.89, | |
| "learning_rate": 3.6115402389912197e-06, | |
| "loss": 0.5153, | |
| "step": 352500 | |
| }, | |
| { | |
| "epoch": 41.95, | |
| "learning_rate": 3.5851323694460956e-06, | |
| "loss": 0.5018, | |
| "step": 353000 | |
| }, | |
| { | |
| "epoch": 42.01, | |
| "learning_rate": 3.5587244999009707e-06, | |
| "loss": 0.5066, | |
| "step": 353500 | |
| }, | |
| { | |
| "epoch": 42.07, | |
| "learning_rate": 3.5323694460949365e-06, | |
| "loss": 0.5071, | |
| "step": 354000 | |
| }, | |
| { | |
| "epoch": 42.13, | |
| "learning_rate": 3.5059615765498124e-06, | |
| "loss": 0.5045, | |
| "step": 354500 | |
| }, | |
| { | |
| "epoch": 42.19, | |
| "learning_rate": 3.4795537070046875e-06, | |
| "loss": 0.4959, | |
| "step": 355000 | |
| }, | |
| { | |
| "epoch": 42.25, | |
| "learning_rate": 3.4531458374595634e-06, | |
| "loss": 0.5034, | |
| "step": 355500 | |
| }, | |
| { | |
| "epoch": 42.31, | |
| "learning_rate": 3.426790783653529e-06, | |
| "loss": 0.5075, | |
| "step": 356000 | |
| }, | |
| { | |
| "epoch": 42.36, | |
| "learning_rate": 3.400435729847495e-06, | |
| "loss": 0.5076, | |
| "step": 356500 | |
| }, | |
| { | |
| "epoch": 42.42, | |
| "learning_rate": 3.37402786030237e-06, | |
| "loss": 0.4914, | |
| "step": 357000 | |
| }, | |
| { | |
| "epoch": 42.48, | |
| "learning_rate": 3.347619990757246e-06, | |
| "loss": 0.4985, | |
| "step": 357500 | |
| }, | |
| { | |
| "epoch": 42.54, | |
| "learning_rate": 3.321212121212121e-06, | |
| "loss": 0.5032, | |
| "step": 358000 | |
| }, | |
| { | |
| "epoch": 42.6, | |
| "learning_rate": 3.294804251666997e-06, | |
| "loss": 0.5022, | |
| "step": 358500 | |
| }, | |
| { | |
| "epoch": 42.66, | |
| "learning_rate": 3.268396382121873e-06, | |
| "loss": 0.5061, | |
| "step": 359000 | |
| }, | |
| { | |
| "epoch": 42.72, | |
| "learning_rate": 3.241988512576748e-06, | |
| "loss": 0.5, | |
| "step": 359500 | |
| }, | |
| { | |
| "epoch": 42.78, | |
| "learning_rate": 3.215580643031624e-06, | |
| "loss": 0.5041, | |
| "step": 360000 | |
| }, | |
| { | |
| "epoch": 42.84, | |
| "learning_rate": 3.189172773486499e-06, | |
| "loss": 0.5044, | |
| "step": 360500 | |
| }, | |
| { | |
| "epoch": 42.9, | |
| "learning_rate": 3.162817719680465e-06, | |
| "loss": 0.502, | |
| "step": 361000 | |
| }, | |
| { | |
| "epoch": 42.96, | |
| "learning_rate": 3.1364098501353407e-06, | |
| "loss": 0.5066, | |
| "step": 361500 | |
| }, | |
| { | |
| "epoch": 43.02, | |
| "learning_rate": 3.110001980590216e-06, | |
| "loss": 0.508, | |
| "step": 362000 | |
| }, | |
| { | |
| "epoch": 43.08, | |
| "learning_rate": 3.0835941110450917e-06, | |
| "loss": 0.4986, | |
| "step": 362500 | |
| }, | |
| { | |
| "epoch": 43.14, | |
| "learning_rate": 3.0571862414999672e-06, | |
| "loss": 0.5001, | |
| "step": 363000 | |
| }, | |
| { | |
| "epoch": 43.2, | |
| "learning_rate": 3.0308840034330233e-06, | |
| "loss": 0.4982, | |
| "step": 363500 | |
| }, | |
| { | |
| "epoch": 43.26, | |
| "learning_rate": 3.004476133887899e-06, | |
| "loss": 0.5028, | |
| "step": 364000 | |
| }, | |
| { | |
| "epoch": 43.32, | |
| "learning_rate": 2.9780682643427743e-06, | |
| "loss": 0.5069, | |
| "step": 364500 | |
| }, | |
| { | |
| "epoch": 43.37, | |
| "learning_rate": 2.9516603947976502e-06, | |
| "loss": 0.5049, | |
| "step": 365000 | |
| }, | |
| { | |
| "epoch": 43.43, | |
| "learning_rate": 2.9252525252525253e-06, | |
| "loss": 0.499, | |
| "step": 365500 | |
| }, | |
| { | |
| "epoch": 43.49, | |
| "learning_rate": 2.8988446557074012e-06, | |
| "loss": 0.5101, | |
| "step": 366000 | |
| }, | |
| { | |
| "epoch": 43.55, | |
| "learning_rate": 2.8724367861622763e-06, | |
| "loss": 0.4984, | |
| "step": 366500 | |
| }, | |
| { | |
| "epoch": 43.61, | |
| "learning_rate": 2.8460289166171522e-06, | |
| "loss": 0.5001, | |
| "step": 367000 | |
| }, | |
| { | |
| "epoch": 43.67, | |
| "learning_rate": 2.8196210470720277e-06, | |
| "loss": 0.504, | |
| "step": 367500 | |
| }, | |
| { | |
| "epoch": 43.73, | |
| "learning_rate": 2.7932659932659935e-06, | |
| "loss": 0.502, | |
| "step": 368000 | |
| }, | |
| { | |
| "epoch": 43.79, | |
| "learning_rate": 2.766858123720869e-06, | |
| "loss": 0.4984, | |
| "step": 368500 | |
| }, | |
| { | |
| "epoch": 43.85, | |
| "learning_rate": 2.7404502541757445e-06, | |
| "loss": 0.5024, | |
| "step": 369000 | |
| }, | |
| { | |
| "epoch": 43.91, | |
| "learning_rate": 2.71404238463062e-06, | |
| "loss": 0.5015, | |
| "step": 369500 | |
| }, | |
| { | |
| "epoch": 43.97, | |
| "learning_rate": 2.687687330824586e-06, | |
| "loss": 0.4996, | |
| "step": 370000 | |
| }, | |
| { | |
| "epoch": 44.03, | |
| "learning_rate": 2.6612794612794613e-06, | |
| "loss": 0.4953, | |
| "step": 370500 | |
| }, | |
| { | |
| "epoch": 44.09, | |
| "learning_rate": 2.634871591734337e-06, | |
| "loss": 0.4966, | |
| "step": 371000 | |
| }, | |
| { | |
| "epoch": 44.15, | |
| "learning_rate": 2.6084637221892128e-06, | |
| "loss": 0.4956, | |
| "step": 371500 | |
| }, | |
| { | |
| "epoch": 44.21, | |
| "learning_rate": 2.5820558526440883e-06, | |
| "loss": 0.4997, | |
| "step": 372000 | |
| }, | |
| { | |
| "epoch": 44.27, | |
| "learning_rate": 2.555700798838054e-06, | |
| "loss": 0.4986, | |
| "step": 372500 | |
| }, | |
| { | |
| "epoch": 44.33, | |
| "learning_rate": 2.5292929292929296e-06, | |
| "loss": 0.502, | |
| "step": 373000 | |
| }, | |
| { | |
| "epoch": 44.39, | |
| "learning_rate": 2.502885059747805e-06, | |
| "loss": 0.4984, | |
| "step": 373500 | |
| }, | |
| { | |
| "epoch": 44.44, | |
| "learning_rate": 2.4764771902026806e-06, | |
| "loss": 0.4984, | |
| "step": 374000 | |
| }, | |
| { | |
| "epoch": 44.5, | |
| "learning_rate": 2.4501221363966464e-06, | |
| "loss": 0.502, | |
| "step": 374500 | |
| }, | |
| { | |
| "epoch": 44.56, | |
| "learning_rate": 2.423714266851522e-06, | |
| "loss": 0.4959, | |
| "step": 375000 | |
| }, | |
| { | |
| "epoch": 44.62, | |
| "learning_rate": 2.3973063973063978e-06, | |
| "loss": 0.4984, | |
| "step": 375500 | |
| }, | |
| { | |
| "epoch": 44.68, | |
| "learning_rate": 2.3708985277612733e-06, | |
| "loss": 0.5018, | |
| "step": 376000 | |
| }, | |
| { | |
| "epoch": 44.74, | |
| "learning_rate": 2.3444906582161488e-06, | |
| "loss": 0.5011, | |
| "step": 376500 | |
| }, | |
| { | |
| "epoch": 44.8, | |
| "learning_rate": 2.3180827886710243e-06, | |
| "loss": 0.503, | |
| "step": 377000 | |
| }, | |
| { | |
| "epoch": 44.86, | |
| "learning_rate": 2.2916749191258998e-06, | |
| "loss": 0.4926, | |
| "step": 377500 | |
| }, | |
| { | |
| "epoch": 44.92, | |
| "learning_rate": 2.2652670495807753e-06, | |
| "loss": 0.4976, | |
| "step": 378000 | |
| }, | |
| { | |
| "epoch": 44.98, | |
| "learning_rate": 2.238911995774741e-06, | |
| "loss": 0.5007, | |
| "step": 378500 | |
| }, | |
| { | |
| "epoch": 45.04, | |
| "learning_rate": 2.2125041262296166e-06, | |
| "loss": 0.5084, | |
| "step": 379000 | |
| }, | |
| { | |
| "epoch": 45.1, | |
| "learning_rate": 2.186096256684492e-06, | |
| "loss": 0.5022, | |
| "step": 379500 | |
| }, | |
| { | |
| "epoch": 45.16, | |
| "learning_rate": 2.1596883871393676e-06, | |
| "loss": 0.4922, | |
| "step": 380000 | |
| }, | |
| { | |
| "epoch": 45.22, | |
| "learning_rate": 2.1332805175942435e-06, | |
| "loss": 0.4957, | |
| "step": 380500 | |
| }, | |
| { | |
| "epoch": 45.28, | |
| "learning_rate": 2.1069254637882093e-06, | |
| "loss": 0.5032, | |
| "step": 381000 | |
| }, | |
| { | |
| "epoch": 45.34, | |
| "learning_rate": 2.080570409982175e-06, | |
| "loss": 0.504, | |
| "step": 381500 | |
| }, | |
| { | |
| "epoch": 45.4, | |
| "learning_rate": 2.0541625404370506e-06, | |
| "loss": 0.5108, | |
| "step": 382000 | |
| }, | |
| { | |
| "epoch": 45.45, | |
| "learning_rate": 2.027754670891926e-06, | |
| "loss": 0.4992, | |
| "step": 382500 | |
| }, | |
| { | |
| "epoch": 45.51, | |
| "learning_rate": 2.0013468013468016e-06, | |
| "loss": 0.4986, | |
| "step": 383000 | |
| }, | |
| { | |
| "epoch": 45.57, | |
| "learning_rate": 1.974938931801677e-06, | |
| "loss": 0.4964, | |
| "step": 383500 | |
| }, | |
| { | |
| "epoch": 45.63, | |
| "learning_rate": 1.948583877995643e-06, | |
| "loss": 0.4974, | |
| "step": 384000 | |
| }, | |
| { | |
| "epoch": 45.69, | |
| "learning_rate": 1.9221760084505184e-06, | |
| "loss": 0.4955, | |
| "step": 384500 | |
| }, | |
| { | |
| "epoch": 45.75, | |
| "learning_rate": 1.895768138905394e-06, | |
| "loss": 0.5006, | |
| "step": 385000 | |
| }, | |
| { | |
| "epoch": 45.81, | |
| "learning_rate": 1.8693602693602694e-06, | |
| "loss": 0.5056, | |
| "step": 385500 | |
| }, | |
| { | |
| "epoch": 45.87, | |
| "learning_rate": 1.842952399815145e-06, | |
| "loss": 0.5018, | |
| "step": 386000 | |
| }, | |
| { | |
| "epoch": 45.93, | |
| "learning_rate": 1.8165445302700204e-06, | |
| "loss": 0.4974, | |
| "step": 386500 | |
| }, | |
| { | |
| "epoch": 45.99, | |
| "learning_rate": 1.7901366607248963e-06, | |
| "loss": 0.5057, | |
| "step": 387000 | |
| }, | |
| { | |
| "epoch": 46.05, | |
| "learning_rate": 1.7637816069188621e-06, | |
| "loss": 0.4892, | |
| "step": 387500 | |
| }, | |
| { | |
| "epoch": 46.11, | |
| "learning_rate": 1.7373737373737376e-06, | |
| "loss": 0.4942, | |
| "step": 388000 | |
| }, | |
| { | |
| "epoch": 46.17, | |
| "learning_rate": 1.7109658678286131e-06, | |
| "loss": 0.5011, | |
| "step": 388500 | |
| }, | |
| { | |
| "epoch": 46.23, | |
| "learning_rate": 1.6845579982834886e-06, | |
| "loss": 0.4945, | |
| "step": 389000 | |
| }, | |
| { | |
| "epoch": 46.29, | |
| "learning_rate": 1.6581501287383641e-06, | |
| "loss": 0.4942, | |
| "step": 389500 | |
| }, | |
| { | |
| "epoch": 46.35, | |
| "learning_rate": 1.6317422591932396e-06, | |
| "loss": 0.4968, | |
| "step": 390000 | |
| }, | |
| { | |
| "epoch": 46.41, | |
| "learning_rate": 1.6053343896481151e-06, | |
| "loss": 0.4979, | |
| "step": 390500 | |
| }, | |
| { | |
| "epoch": 46.46, | |
| "learning_rate": 1.5789265201029908e-06, | |
| "loss": 0.5005, | |
| "step": 391000 | |
| }, | |
| { | |
| "epoch": 46.52, | |
| "learning_rate": 1.5525186505578663e-06, | |
| "loss": 0.4946, | |
| "step": 391500 | |
| }, | |
| { | |
| "epoch": 46.58, | |
| "learning_rate": 1.5261107810127418e-06, | |
| "loss": 0.4923, | |
| "step": 392000 | |
| }, | |
| { | |
| "epoch": 46.64, | |
| "learning_rate": 1.4997557272067076e-06, | |
| "loss": 0.5, | |
| "step": 392500 | |
| }, | |
| { | |
| "epoch": 46.7, | |
| "learning_rate": 1.4733478576615833e-06, | |
| "loss": 0.4971, | |
| "step": 393000 | |
| }, | |
| { | |
| "epoch": 46.76, | |
| "learning_rate": 1.4469399881164588e-06, | |
| "loss": 0.4977, | |
| "step": 393500 | |
| }, | |
| { | |
| "epoch": 46.82, | |
| "learning_rate": 1.4205321185713343e-06, | |
| "loss": 0.4925, | |
| "step": 394000 | |
| }, | |
| { | |
| "epoch": 46.88, | |
| "learning_rate": 1.3941770647653001e-06, | |
| "loss": 0.4987, | |
| "step": 394500 | |
| }, | |
| { | |
| "epoch": 46.94, | |
| "learning_rate": 1.3677691952201759e-06, | |
| "loss": 0.5004, | |
| "step": 395000 | |
| }, | |
| { | |
| "epoch": 47.0, | |
| "learning_rate": 1.3413613256750514e-06, | |
| "loss": 0.4961, | |
| "step": 395500 | |
| }, | |
| { | |
| "epoch": 47.06, | |
| "learning_rate": 1.3149534561299269e-06, | |
| "loss": 0.494, | |
| "step": 396000 | |
| }, | |
| { | |
| "epoch": 47.12, | |
| "learning_rate": 1.2885455865848024e-06, | |
| "loss": 0.4941, | |
| "step": 396500 | |
| }, | |
| { | |
| "epoch": 47.18, | |
| "learning_rate": 1.2621377170396779e-06, | |
| "loss": 0.5015, | |
| "step": 397000 | |
| }, | |
| { | |
| "epoch": 47.24, | |
| "learning_rate": 1.2357298474945536e-06, | |
| "loss": 0.4976, | |
| "step": 397500 | |
| }, | |
| { | |
| "epoch": 47.3, | |
| "learning_rate": 1.209321977949429e-06, | |
| "loss": 0.4887, | |
| "step": 398000 | |
| }, | |
| { | |
| "epoch": 47.36, | |
| "learning_rate": 1.1829141084043046e-06, | |
| "loss": 0.5032, | |
| "step": 398500 | |
| }, | |
| { | |
| "epoch": 47.42, | |
| "learning_rate": 1.15650623885918e-06, | |
| "loss": 0.4959, | |
| "step": 399000 | |
| }, | |
| { | |
| "epoch": 47.47, | |
| "learning_rate": 1.1302568165313265e-06, | |
| "loss": 0.4965, | |
| "step": 399500 | |
| }, | |
| { | |
| "epoch": 47.53, | |
| "learning_rate": 1.103848946986202e-06, | |
| "loss": 0.4949, | |
| "step": 400000 | |
| }, | |
| { | |
| "epoch": 47.59, | |
| "learning_rate": 1.0774938931801678e-06, | |
| "loss": 0.4879, | |
| "step": 400500 | |
| }, | |
| { | |
| "epoch": 47.65, | |
| "learning_rate": 1.0510860236350433e-06, | |
| "loss": 0.4935, | |
| "step": 401000 | |
| }, | |
| { | |
| "epoch": 47.71, | |
| "learning_rate": 1.024678154089919e-06, | |
| "loss": 0.4954, | |
| "step": 401500 | |
| }, | |
| { | |
| "epoch": 47.77, | |
| "learning_rate": 9.982702845447945e-07, | |
| "loss": 0.4919, | |
| "step": 402000 | |
| }, | |
| { | |
| "epoch": 47.83, | |
| "learning_rate": 9.7186241499967e-07, | |
| "loss": 0.4974, | |
| "step": 402500 | |
| }, | |
| { | |
| "epoch": 47.89, | |
| "learning_rate": 9.454545454545455e-07, | |
| "loss": 0.5004, | |
| "step": 403000 | |
| }, | |
| { | |
| "epoch": 47.95, | |
| "learning_rate": 9.190466759094211e-07, | |
| "loss": 0.5003, | |
| "step": 403500 | |
| }, | |
| { | |
| "epoch": 48.01, | |
| "learning_rate": 8.926388063642967e-07, | |
| "loss": 0.5044, | |
| "step": 404000 | |
| }, | |
| { | |
| "epoch": 48.07, | |
| "learning_rate": 8.662309368191722e-07, | |
| "loss": 0.489, | |
| "step": 404500 | |
| }, | |
| { | |
| "epoch": 48.13, | |
| "learning_rate": 8.398230672740478e-07, | |
| "loss": 0.4942, | |
| "step": 405000 | |
| }, | |
| { | |
| "epoch": 48.19, | |
| "learning_rate": 8.134151977289233e-07, | |
| "loss": 0.4996, | |
| "step": 405500 | |
| }, | |
| { | |
| "epoch": 48.25, | |
| "learning_rate": 7.870073281837988e-07, | |
| "loss": 0.4896, | |
| "step": 406000 | |
| }, | |
| { | |
| "epoch": 48.31, | |
| "learning_rate": 7.605994586386743e-07, | |
| "loss": 0.494, | |
| "step": 406500 | |
| }, | |
| { | |
| "epoch": 48.37, | |
| "learning_rate": 7.3419158909355e-07, | |
| "loss": 0.4911, | |
| "step": 407000 | |
| }, | |
| { | |
| "epoch": 48.43, | |
| "learning_rate": 7.077837195484255e-07, | |
| "loss": 0.4961, | |
| "step": 407500 | |
| }, | |
| { | |
| "epoch": 48.48, | |
| "learning_rate": 6.81375850003301e-07, | |
| "loss": 0.4894, | |
| "step": 408000 | |
| }, | |
| { | |
| "epoch": 48.54, | |
| "learning_rate": 6.549679804581765e-07, | |
| "loss": 0.4952, | |
| "step": 408500 | |
| }, | |
| { | |
| "epoch": 48.6, | |
| "learning_rate": 6.285601109130521e-07, | |
| "loss": 0.5021, | |
| "step": 409000 | |
| }, | |
| { | |
| "epoch": 48.66, | |
| "learning_rate": 6.021522413679277e-07, | |
| "loss": 0.5009, | |
| "step": 409500 | |
| }, | |
| { | |
| "epoch": 48.72, | |
| "learning_rate": 5.757971875618935e-07, | |
| "loss": 0.4962, | |
| "step": 410000 | |
| }, | |
| { | |
| "epoch": 48.78, | |
| "learning_rate": 5.49389318016769e-07, | |
| "loss": 0.4939, | |
| "step": 410500 | |
| }, | |
| { | |
| "epoch": 48.84, | |
| "learning_rate": 5.229814484716446e-07, | |
| "loss": 0.497, | |
| "step": 411000 | |
| }, | |
| { | |
| "epoch": 48.9, | |
| "learning_rate": 4.965735789265201e-07, | |
| "loss": 0.4995, | |
| "step": 411500 | |
| }, | |
| { | |
| "epoch": 48.96, | |
| "learning_rate": 4.7016570938139573e-07, | |
| "loss": 0.5008, | |
| "step": 412000 | |
| }, | |
| { | |
| "epoch": 49.02, | |
| "learning_rate": 4.4386347131445177e-07, | |
| "loss": 0.5001, | |
| "step": 412500 | |
| }, | |
| { | |
| "epoch": 49.08, | |
| "learning_rate": 4.174556017693273e-07, | |
| "loss": 0.4934, | |
| "step": 413000 | |
| }, | |
| { | |
| "epoch": 49.14, | |
| "learning_rate": 3.910477322242028e-07, | |
| "loss": 0.4949, | |
| "step": 413500 | |
| }, | |
| { | |
| "epoch": 49.2, | |
| "learning_rate": 3.6463986267907843e-07, | |
| "loss": 0.5009, | |
| "step": 414000 | |
| }, | |
| { | |
| "epoch": 49.26, | |
| "learning_rate": 3.3823199313395393e-07, | |
| "loss": 0.5013, | |
| "step": 414500 | |
| }, | |
| { | |
| "epoch": 49.32, | |
| "learning_rate": 3.118241235888295e-07, | |
| "loss": 0.5009, | |
| "step": 415000 | |
| }, | |
| { | |
| "epoch": 49.38, | |
| "learning_rate": 2.8541625404370504e-07, | |
| "loss": 0.4927, | |
| "step": 415500 | |
| }, | |
| { | |
| "epoch": 49.44, | |
| "learning_rate": 2.590083844985806e-07, | |
| "loss": 0.5005, | |
| "step": 416000 | |
| }, | |
| { | |
| "epoch": 49.49, | |
| "learning_rate": 2.326533306925464e-07, | |
| "loss": 0.4908, | |
| "step": 416500 | |
| }, | |
| { | |
| "epoch": 49.55, | |
| "learning_rate": 2.0624546114742194e-07, | |
| "loss": 0.4904, | |
| "step": 417000 | |
| }, | |
| { | |
| "epoch": 49.61, | |
| "learning_rate": 1.798375916022975e-07, | |
| "loss": 0.5013, | |
| "step": 417500 | |
| }, | |
| { | |
| "epoch": 49.67, | |
| "learning_rate": 1.5342972205717305e-07, | |
| "loss": 0.4991, | |
| "step": 418000 | |
| }, | |
| { | |
| "epoch": 49.73, | |
| "learning_rate": 1.270218525120486e-07, | |
| "loss": 0.4921, | |
| "step": 418500 | |
| }, | |
| { | |
| "epoch": 49.79, | |
| "learning_rate": 1.006667987060144e-07, | |
| "loss": 0.4987, | |
| "step": 419000 | |
| }, | |
| { | |
| "epoch": 49.85, | |
| "learning_rate": 7.425892916088995e-08, | |
| "loss": 0.4946, | |
| "step": 419500 | |
| }, | |
| { | |
| "epoch": 49.91, | |
| "learning_rate": 4.790387535485575e-08, | |
| "loss": 0.492, | |
| "step": 420000 | |
| }, | |
| { | |
| "epoch": 49.97, | |
| "learning_rate": 2.14960058097313e-08, | |
| "loss": 0.4974, | |
| "step": 420500 | |
| } | |
| ], | |
| "max_steps": 420750, | |
| "num_train_epochs": 50, | |
| "total_flos": 7.065593235141296e+19, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |