{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 585, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0, "forget_Q_A_ROUGE": 0.3188276274014988, "step": 0 }, { "epoch": 0.042735042735042736, "grad_norm": 764.0, "learning_rate": 4.3103448275862073e-07, "loss": 14.2, "step": 5 }, { "epoch": 0.08547008547008547, "grad_norm": 652.0, "learning_rate": 8.620689655172415e-07, "loss": 14.0957, "step": 10 }, { "epoch": 0.1282051282051282, "grad_norm": 700.0, "learning_rate": 1.2931034482758623e-06, "loss": 13.9925, "step": 15 }, { "epoch": 0.17094017094017094, "grad_norm": 740.0, "learning_rate": 1.724137931034483e-06, "loss": 13.8799, "step": 20 }, { "epoch": 0.21367521367521367, "grad_norm": 692.0, "learning_rate": 2.1551724137931035e-06, "loss": 13.7433, "step": 25 }, { "epoch": 0.2564102564102564, "grad_norm": 608.0, "learning_rate": 2.5862068965517246e-06, "loss": 13.5505, "step": 30 }, { "epoch": 0.29914529914529914, "grad_norm": 676.0, "learning_rate": 3.017241379310345e-06, "loss": 13.3642, "step": 35 }, { "epoch": 0.3418803418803419, "grad_norm": 760.0, "learning_rate": 3.448275862068966e-06, "loss": 12.9769, "step": 40 }, { "epoch": 0.38461538461538464, "grad_norm": 728.0, "learning_rate": 3.8793103448275865e-06, "loss": 11.7681, "step": 45 }, { "epoch": 0.42735042735042733, "grad_norm": 728.0, "learning_rate": 4.310344827586207e-06, "loss": 10.5377, "step": 50 }, { "epoch": 0.4700854700854701, "grad_norm": 676.0, "learning_rate": 4.741379310344828e-06, "loss": 8.6128, "step": 55 }, { "epoch": 0.5128205128205128, "grad_norm": 664.0, "learning_rate": 5.172413793103449e-06, "loss": 6.0024, "step": 60 }, { "epoch": 0.5555555555555556, "grad_norm": 312.0, "learning_rate": 5.603448275862069e-06, "loss": 2.7774, "step": 65 }, { "epoch": 0.5982905982905983, "grad_norm": 96.5, "learning_rate": 6.03448275862069e-06, "loss": 1.5545, "step": 70 }, { "epoch": 0.6410256410256411, "grad_norm": 79.0, "learning_rate": 6.465517241379311e-06, "loss": 1.4243, "step": 75 }, { "epoch": 0.6837606837606838, "grad_norm": 77.0, "learning_rate": 6.896551724137932e-06, "loss": 1.159, "step": 80 }, { "epoch": 0.7264957264957265, "grad_norm": 81.0, "learning_rate": 7.327586206896552e-06, "loss": 1.0177, "step": 85 }, { "epoch": 0.7692307692307693, "grad_norm": 46.25, "learning_rate": 7.758620689655173e-06, "loss": 0.849, "step": 90 }, { "epoch": 0.811965811965812, "grad_norm": 31.375, "learning_rate": 8.189655172413794e-06, "loss": 0.816, "step": 95 }, { "epoch": 0.8547008547008547, "grad_norm": 39.5, "learning_rate": 8.620689655172414e-06, "loss": 0.728, "step": 100 }, { "epoch": 0.8974358974358975, "grad_norm": 99.0, "learning_rate": 9.051724137931036e-06, "loss": 0.7868, "step": 105 }, { "epoch": 0.9401709401709402, "grad_norm": 26.875, "learning_rate": 9.482758620689655e-06, "loss": 0.5595, "step": 110 }, { "epoch": 0.9829059829059829, "grad_norm": 22.625, "learning_rate": 9.913793103448277e-06, "loss": 0.5535, "step": 115 }, { "epoch": 1.0, "forget_Q_A_ROUGE": 0.18181675542335068, "step": 117 }, { "epoch": 1.0256410256410255, "grad_norm": 17.75, "learning_rate": 9.914712153518125e-06, "loss": 0.5077, "step": 120 }, { "epoch": 1.0683760683760684, "grad_norm": 12.125, "learning_rate": 9.80810234541578e-06, "loss": 0.4655, "step": 125 }, { "epoch": 1.1111111111111112, "grad_norm": 31.25, "learning_rate": 9.701492537313434e-06, "loss": 0.4229, "step": 130 }, { "epoch": 1.1538461538461537, "grad_norm": 24.125, "learning_rate": 9.594882729211089e-06, "loss": 0.4191, "step": 135 }, { "epoch": 1.1965811965811965, "grad_norm": 21.25, "learning_rate": 9.488272921108744e-06, "loss": 0.3993, "step": 140 }, { "epoch": 1.2393162393162394, "grad_norm": 10.75, "learning_rate": 9.381663113006397e-06, "loss": 0.3551, "step": 145 }, { "epoch": 1.282051282051282, "grad_norm": 7.71875, "learning_rate": 9.275053304904051e-06, "loss": 0.3124, "step": 150 }, { "epoch": 1.3247863247863247, "grad_norm": 8.0625, "learning_rate": 9.168443496801706e-06, "loss": 0.3136, "step": 155 }, { "epoch": 1.3675213675213675, "grad_norm": 15.4375, "learning_rate": 9.06183368869936e-06, "loss": 0.3554, "step": 160 }, { "epoch": 1.4102564102564101, "grad_norm": 7.46875, "learning_rate": 8.955223880597016e-06, "loss": 0.2964, "step": 165 }, { "epoch": 1.452991452991453, "grad_norm": 62.0, "learning_rate": 8.84861407249467e-06, "loss": 0.2874, "step": 170 }, { "epoch": 1.4957264957264957, "grad_norm": 13.9375, "learning_rate": 8.742004264392325e-06, "loss": 0.2957, "step": 175 }, { "epoch": 1.5384615384615383, "grad_norm": 8.125, "learning_rate": 8.63539445628998e-06, "loss": 0.3001, "step": 180 }, { "epoch": 1.5811965811965814, "grad_norm": 6.625, "learning_rate": 8.528784648187633e-06, "loss": 0.3171, "step": 185 }, { "epoch": 1.623931623931624, "grad_norm": 6.0625, "learning_rate": 8.42217484008529e-06, "loss": 0.2736, "step": 190 }, { "epoch": 1.6666666666666665, "grad_norm": 7.59375, "learning_rate": 8.315565031982942e-06, "loss": 0.3145, "step": 195 }, { "epoch": 1.7094017094017095, "grad_norm": 13.4375, "learning_rate": 8.208955223880599e-06, "loss": 0.2538, "step": 200 }, { "epoch": 1.7521367521367521, "grad_norm": 18.75, "learning_rate": 8.102345415778252e-06, "loss": 0.2419, "step": 205 }, { "epoch": 1.7948717948717947, "grad_norm": 7.09375, "learning_rate": 7.995735607675907e-06, "loss": 0.2446, "step": 210 }, { "epoch": 1.8376068376068377, "grad_norm": 5.5625, "learning_rate": 7.889125799573561e-06, "loss": 0.2172, "step": 215 }, { "epoch": 1.8803418803418803, "grad_norm": 6.75, "learning_rate": 7.782515991471216e-06, "loss": 0.2354, "step": 220 }, { "epoch": 1.9230769230769231, "grad_norm": 6.28125, "learning_rate": 7.67590618336887e-06, "loss": 0.2031, "step": 225 }, { "epoch": 1.965811965811966, "grad_norm": 7.09375, "learning_rate": 7.569296375266525e-06, "loss": 0.2396, "step": 230 }, { "epoch": 2.0, "forget_Q_A_ROUGE": 0.16664429628381783, "step": 234 }, { "epoch": 2.0085470085470085, "grad_norm": 7.03125, "learning_rate": 7.46268656716418e-06, "loss": 0.2064, "step": 235 }, { "epoch": 2.051282051282051, "grad_norm": 5.5625, "learning_rate": 7.356076759061834e-06, "loss": 0.1929, "step": 240 }, { "epoch": 2.094017094017094, "grad_norm": 6.4375, "learning_rate": 7.249466950959488e-06, "loss": 0.1718, "step": 245 }, { "epoch": 2.1367521367521367, "grad_norm": 6.1875, "learning_rate": 7.1428571428571436e-06, "loss": 0.2016, "step": 250 }, { "epoch": 2.1794871794871793, "grad_norm": 6.0, "learning_rate": 7.0362473347547975e-06, "loss": 0.1856, "step": 255 }, { "epoch": 2.2222222222222223, "grad_norm": 5.0625, "learning_rate": 6.929637526652453e-06, "loss": 0.1694, "step": 260 }, { "epoch": 2.264957264957265, "grad_norm": 5.40625, "learning_rate": 6.823027718550107e-06, "loss": 0.1571, "step": 265 }, { "epoch": 2.3076923076923075, "grad_norm": 7.40625, "learning_rate": 6.7164179104477625e-06, "loss": 0.1763, "step": 270 }, { "epoch": 2.3504273504273505, "grad_norm": 5.5, "learning_rate": 6.609808102345416e-06, "loss": 0.1777, "step": 275 }, { "epoch": 2.393162393162393, "grad_norm": 6.0, "learning_rate": 6.50319829424307e-06, "loss": 0.1754, "step": 280 }, { "epoch": 2.435897435897436, "grad_norm": 5.8125, "learning_rate": 6.396588486140726e-06, "loss": 0.158, "step": 285 }, { "epoch": 2.4786324786324787, "grad_norm": 6.59375, "learning_rate": 6.28997867803838e-06, "loss": 0.1666, "step": 290 }, { "epoch": 2.5213675213675213, "grad_norm": 5.28125, "learning_rate": 6.183368869936035e-06, "loss": 0.1561, "step": 295 }, { "epoch": 2.564102564102564, "grad_norm": 5.25, "learning_rate": 6.076759061833689e-06, "loss": 0.1373, "step": 300 }, { "epoch": 2.606837606837607, "grad_norm": 6.46875, "learning_rate": 5.970149253731343e-06, "loss": 0.1559, "step": 305 }, { "epoch": 2.6495726495726495, "grad_norm": 5.90625, "learning_rate": 5.863539445628999e-06, "loss": 0.1453, "step": 310 }, { "epoch": 2.6923076923076925, "grad_norm": 4.9375, "learning_rate": 5.756929637526653e-06, "loss": 0.153, "step": 315 }, { "epoch": 2.735042735042735, "grad_norm": 5.5, "learning_rate": 5.650319829424308e-06, "loss": 0.1456, "step": 320 }, { "epoch": 2.7777777777777777, "grad_norm": 6.6875, "learning_rate": 5.543710021321962e-06, "loss": 0.1375, "step": 325 }, { "epoch": 2.8205128205128203, "grad_norm": 5.125, "learning_rate": 5.437100213219617e-06, "loss": 0.1433, "step": 330 }, { "epoch": 2.8632478632478633, "grad_norm": 6.34375, "learning_rate": 5.3304904051172716e-06, "loss": 0.1354, "step": 335 }, { "epoch": 2.905982905982906, "grad_norm": 7.5, "learning_rate": 5.2238805970149255e-06, "loss": 0.1252, "step": 340 }, { "epoch": 2.948717948717949, "grad_norm": 5.15625, "learning_rate": 5.11727078891258e-06, "loss": 0.1273, "step": 345 }, { "epoch": 2.9914529914529915, "grad_norm": 4.53125, "learning_rate": 5.010660980810235e-06, "loss": 0.1179, "step": 350 }, { "epoch": 3.0, "forget_Q_A_ROUGE": 0.16380368545856266, "step": 351 }, { "epoch": 3.034188034188034, "grad_norm": 5.4375, "learning_rate": 4.90405117270789e-06, "loss": 0.1379, "step": 355 }, { "epoch": 3.076923076923077, "grad_norm": 5.34375, "learning_rate": 4.797441364605544e-06, "loss": 0.1277, "step": 360 }, { "epoch": 3.1196581196581197, "grad_norm": 5.0, "learning_rate": 4.690831556503198e-06, "loss": 0.1305, "step": 365 }, { "epoch": 3.1623931623931623, "grad_norm": 4.71875, "learning_rate": 4.584221748400853e-06, "loss": 0.118, "step": 370 }, { "epoch": 3.2051282051282053, "grad_norm": 4.84375, "learning_rate": 4.477611940298508e-06, "loss": 0.1176, "step": 375 }, { "epoch": 3.247863247863248, "grad_norm": 4.75, "learning_rate": 4.3710021321961625e-06, "loss": 0.12, "step": 380 }, { "epoch": 3.2905982905982905, "grad_norm": 4.34375, "learning_rate": 4.264392324093816e-06, "loss": 0.1288, "step": 385 }, { "epoch": 3.3333333333333335, "grad_norm": 5.34375, "learning_rate": 4.157782515991471e-06, "loss": 0.1244, "step": 390 }, { "epoch": 3.376068376068376, "grad_norm": 4.4375, "learning_rate": 4.051172707889126e-06, "loss": 0.1189, "step": 395 }, { "epoch": 3.4188034188034186, "grad_norm": 4.71875, "learning_rate": 3.944562899786781e-06, "loss": 0.1073, "step": 400 }, { "epoch": 3.4615384615384617, "grad_norm": 4.6875, "learning_rate": 3.837953091684435e-06, "loss": 0.1056, "step": 405 }, { "epoch": 3.5042735042735043, "grad_norm": 3.96875, "learning_rate": 3.73134328358209e-06, "loss": 0.0968, "step": 410 }, { "epoch": 3.547008547008547, "grad_norm": 4.40625, "learning_rate": 3.624733475479744e-06, "loss": 0.114, "step": 415 }, { "epoch": 3.58974358974359, "grad_norm": 4.34375, "learning_rate": 3.5181236673773987e-06, "loss": 0.0984, "step": 420 }, { "epoch": 3.6324786324786325, "grad_norm": 4.0625, "learning_rate": 3.4115138592750535e-06, "loss": 0.1039, "step": 425 }, { "epoch": 3.6752136752136755, "grad_norm": 5.15625, "learning_rate": 3.304904051172708e-06, "loss": 0.1201, "step": 430 }, { "epoch": 3.717948717948718, "grad_norm": 4.78125, "learning_rate": 3.198294243070363e-06, "loss": 0.1138, "step": 435 }, { "epoch": 3.7606837606837606, "grad_norm": 4.4375, "learning_rate": 3.0916844349680177e-06, "loss": 0.1009, "step": 440 }, { "epoch": 3.8034188034188032, "grad_norm": 5.0, "learning_rate": 2.9850746268656716e-06, "loss": 0.1073, "step": 445 }, { "epoch": 3.8461538461538463, "grad_norm": 3.890625, "learning_rate": 2.8784648187633263e-06, "loss": 0.1065, "step": 450 }, { "epoch": 3.888888888888889, "grad_norm": 5.65625, "learning_rate": 2.771855010660981e-06, "loss": 0.1164, "step": 455 }, { "epoch": 3.931623931623932, "grad_norm": 4.0625, "learning_rate": 2.6652452025586358e-06, "loss": 0.0976, "step": 460 }, { "epoch": 3.9743589743589745, "grad_norm": 4.375, "learning_rate": 2.55863539445629e-06, "loss": 0.1027, "step": 465 }, { "epoch": 4.0, "forget_Q_A_ROUGE": 0.1577463934994216, "step": 468 }, { "epoch": 4.017094017094017, "grad_norm": 5.03125, "learning_rate": 2.452025586353945e-06, "loss": 0.108, "step": 470 }, { "epoch": 4.05982905982906, "grad_norm": 3.53125, "learning_rate": 2.345415778251599e-06, "loss": 0.0954, "step": 475 }, { "epoch": 4.102564102564102, "grad_norm": 4.53125, "learning_rate": 2.238805970149254e-06, "loss": 0.107, "step": 480 }, { "epoch": 4.145299145299146, "grad_norm": 4.625, "learning_rate": 2.132196162046908e-06, "loss": 0.1001, "step": 485 }, { "epoch": 4.188034188034188, "grad_norm": 5.28125, "learning_rate": 2.025586353944563e-06, "loss": 0.109, "step": 490 }, { "epoch": 4.230769230769231, "grad_norm": 4.4375, "learning_rate": 1.9189765458422177e-06, "loss": 0.1133, "step": 495 }, { "epoch": 4.273504273504273, "grad_norm": 4.0, "learning_rate": 1.812366737739872e-06, "loss": 0.092, "step": 500 }, { "epoch": 4.316239316239316, "grad_norm": 4.34375, "learning_rate": 1.7057569296375267e-06, "loss": 0.1034, "step": 505 }, { "epoch": 4.358974358974359, "grad_norm": 4.40625, "learning_rate": 1.5991471215351815e-06, "loss": 0.0986, "step": 510 }, { "epoch": 4.401709401709402, "grad_norm": 4.09375, "learning_rate": 1.4925373134328358e-06, "loss": 0.1168, "step": 515 }, { "epoch": 4.444444444444445, "grad_norm": 4.90625, "learning_rate": 1.3859275053304905e-06, "loss": 0.1033, "step": 520 }, { "epoch": 4.487179487179487, "grad_norm": 4.65625, "learning_rate": 1.279317697228145e-06, "loss": 0.1006, "step": 525 }, { "epoch": 4.52991452991453, "grad_norm": 5.21875, "learning_rate": 1.1727078891257996e-06, "loss": 0.109, "step": 530 }, { "epoch": 4.572649572649572, "grad_norm": 4.40625, "learning_rate": 1.066098081023454e-06, "loss": 0.1129, "step": 535 }, { "epoch": 4.615384615384615, "grad_norm": 5.34375, "learning_rate": 9.594882729211088e-07, "loss": 0.0991, "step": 540 }, { "epoch": 4.6581196581196584, "grad_norm": 4.84375, "learning_rate": 8.528784648187634e-07, "loss": 0.1047, "step": 545 }, { "epoch": 4.700854700854701, "grad_norm": 5.59375, "learning_rate": 7.462686567164179e-07, "loss": 0.1099, "step": 550 }, { "epoch": 4.743589743589744, "grad_norm": 3.984375, "learning_rate": 6.396588486140725e-07, "loss": 0.0981, "step": 555 }, { "epoch": 4.786324786324786, "grad_norm": 3.9375, "learning_rate": 5.33049040511727e-07, "loss": 0.0975, "step": 560 }, { "epoch": 4.829059829059829, "grad_norm": 5.40625, "learning_rate": 4.264392324093817e-07, "loss": 0.1066, "step": 565 }, { "epoch": 4.871794871794872, "grad_norm": 4.53125, "learning_rate": 3.1982942430703626e-07, "loss": 0.1064, "step": 570 }, { "epoch": 4.914529914529915, "grad_norm": 3.953125, "learning_rate": 2.1321961620469084e-07, "loss": 0.0975, "step": 575 }, { "epoch": 4.957264957264957, "grad_norm": 3.96875, "learning_rate": 1.0660980810234542e-07, "loss": 0.1096, "step": 580 }, { "epoch": 5.0, "grad_norm": 4.15625, "learning_rate": 0.0, "loss": 0.0957, "step": 585 }, { "epoch": 5.0, "forget_Q_A_ROUGE": 0.16323620599043673, "step": 585 }, { "epoch": 5.0, "step": 585, "total_flos": 0.0, "train_loss": 1.4970051565231421, "train_runtime": 6804.7106, "train_samples_per_second": 5.499, "train_steps_per_second": 0.086 } ], "logging_steps": 5, "max_steps": 585, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }