Amber
Upload checkpoint-19396
405a042
Invalid JSON: Unexpected token 'N', ..."on_loss": NaN, "... is not valid JSON
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9999742221534813,
"eval_steps": 500,
"global_step": 19396,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"decision_loss": 17.2946,
"encoder_mlm_loss": 1.3309,
"epoch": 0.0,
"grad_norm": 506.80633544921875,
"learning_rate": 4.4e-07,
"loss": 37.9076,
"reason_loss": NaN,
"step": 50
},
{
"decision_loss": 15.157,
"encoder_mlm_loss": 1.3116,
"epoch": 0.01,
"grad_norm": 324.48309326171875,
"learning_rate": 9.400000000000001e-07,
"loss": 33.6669,
"reason_loss": NaN,
"step": 100
},
{
"decision_loss": 11.6826,
"encoder_mlm_loss": 1.2069,
"epoch": 0.01,
"grad_norm": 171.52377319335938,
"learning_rate": 1.44e-06,
"loss": 26.0507,
"reason_loss": NaN,
"step": 150
},
{
"decision_loss": 8.4601,
"encoder_mlm_loss": 1.1887,
"epoch": 0.01,
"grad_norm": 69.40190887451172,
"learning_rate": 1.93e-06,
"loss": 19.9829,
"reason_loss": NaN,
"step": 200
},
{
"decision_loss": 6.6507,
"encoder_mlm_loss": 1.0606,
"epoch": 0.01,
"grad_norm": 32.03302764892578,
"learning_rate": 2.43e-06,
"loss": 16.3406,
"reason_loss": NaN,
"step": 250
},
{
"decision_loss": 5.7539,
"encoder_mlm_loss": 0.9811,
"epoch": 0.02,
"grad_norm": 21.778579711914062,
"learning_rate": 2.93e-06,
"loss": 14.2326,
"reason_loss": NaN,
"step": 300
},
{
"decision_loss": 5.3877,
"encoder_mlm_loss": 0.9975,
"epoch": 0.02,
"grad_norm": 18.334007263183594,
"learning_rate": 3.4200000000000003e-06,
"loss": 13.4659,
"reason_loss": NaN,
"step": 350
},
{
"decision_loss": 5.1198,
"encoder_mlm_loss": 0.9698,
"epoch": 0.02,
"grad_norm": 20.7861270904541,
"learning_rate": 3.92e-06,
"loss": 12.9997,
"reason_loss": NaN,
"step": 400
},
{
"decision_loss": 4.9401,
"encoder_mlm_loss": 0.8893,
"epoch": 0.02,
"grad_norm": 16.364755630493164,
"learning_rate": 4.420000000000001e-06,
"loss": 12.7823,
"reason_loss": NaN,
"step": 450
},
{
"decision_loss": 4.8397,
"encoder_mlm_loss": 0.8811,
"epoch": 0.03,
"grad_norm": 15.240585327148438,
"learning_rate": 4.92e-06,
"loss": 12.3987,
"reason_loss": NaN,
"step": 500
},
{
"decision_loss": 4.8283,
"encoder_mlm_loss": 0.8613,
"epoch": 0.03,
"grad_norm": 16.48607063293457,
"learning_rate": 5.42e-06,
"loss": 12.1919,
"reason_loss": NaN,
"step": 550
},
{
"decision_loss": 4.7769,
"encoder_mlm_loss": 0.8291,
"epoch": 0.03,
"grad_norm": 14.643627166748047,
"learning_rate": 5.920000000000001e-06,
"loss": 12.0182,
"reason_loss": NaN,
"step": 600
},
{
"decision_loss": 4.8057,
"encoder_mlm_loss": 0.8424,
"epoch": 0.03,
"grad_norm": 14.431436538696289,
"learning_rate": 6.4199999999999995e-06,
"loss": 11.9925,
"reason_loss": NaN,
"step": 650
},
{
"decision_loss": 4.6025,
"encoder_mlm_loss": 0.8223,
"epoch": 0.04,
"grad_norm": 27.42824935913086,
"learning_rate": 6.92e-06,
"loss": 11.7964,
"reason_loss": NaN,
"step": 700
},
{
"decision_loss": 4.5554,
"encoder_mlm_loss": 0.8215,
"epoch": 0.04,
"grad_norm": 15.171280860900879,
"learning_rate": 7.420000000000001e-06,
"loss": 11.8384,
"reason_loss": NaN,
"step": 750
},
{
"decision_loss": 4.4705,
"encoder_mlm_loss": 0.7881,
"epoch": 0.04,
"grad_norm": 15.859951972961426,
"learning_rate": 7.92e-06,
"loss": 11.9251,
"reason_loss": NaN,
"step": 800
},
{
"decision_loss": 4.502,
"encoder_mlm_loss": 0.7824,
"epoch": 0.04,
"grad_norm": 13.420965194702148,
"learning_rate": 8.42e-06,
"loss": 11.589,
"reason_loss": NaN,
"step": 850
},
{
"decision_loss": 4.4354,
"encoder_mlm_loss": 0.7933,
"epoch": 0.05,
"grad_norm": 20.237834930419922,
"learning_rate": 8.920000000000001e-06,
"loss": 11.4423,
"reason_loss": NaN,
"step": 900
},
{
"decision_loss": 4.4066,
"encoder_mlm_loss": 0.8067,
"epoch": 0.05,
"grad_norm": 14.367069244384766,
"learning_rate": 9.420000000000001e-06,
"loss": 11.5077,
"reason_loss": NaN,
"step": 950
},
{
"decision_loss": 4.4373,
"encoder_mlm_loss": 0.7737,
"epoch": 0.05,
"grad_norm": 14.129799842834473,
"learning_rate": 9.92e-06,
"loss": 11.2995,
"reason_loss": NaN,
"step": 1000
},
{
"decision_loss": 4.374,
"encoder_mlm_loss": 0.7633,
"epoch": 0.05,
"grad_norm": 14.180242538452148,
"learning_rate": 1.042e-05,
"loss": 11.1966,
"reason_loss": 5.5783,
"step": 1050
},
{
"decision_loss": 4.4494,
"encoder_mlm_loss": 0.7522,
"epoch": 0.06,
"grad_norm": 11.483891487121582,
"learning_rate": 1.092e-05,
"loss": 11.5075,
"reason_loss": NaN,
"step": 1100
},
{
"decision_loss": 4.3591,
"encoder_mlm_loss": 0.7654,
"epoch": 0.06,
"grad_norm": 12.64001178741455,
"learning_rate": 1.142e-05,
"loss": 11.0272,
"reason_loss": NaN,
"step": 1150
},
{
"decision_loss": 4.3741,
"encoder_mlm_loss": 0.7449,
"epoch": 0.06,
"grad_norm": 14.490357398986816,
"learning_rate": 1.1920000000000001e-05,
"loss": 11.2744,
"reason_loss": NaN,
"step": 1200
},
{
"decision_loss": 4.3079,
"encoder_mlm_loss": 0.724,
"epoch": 0.06,
"grad_norm": 12.611394882202148,
"learning_rate": 1.2420000000000001e-05,
"loss": 11.3553,
"reason_loss": NaN,
"step": 1250
},
{
"decision_loss": 4.331,
"encoder_mlm_loss": 0.7394,
"epoch": 0.07,
"grad_norm": 23.937461853027344,
"learning_rate": 1.2920000000000002e-05,
"loss": 11.4936,
"reason_loss": NaN,
"step": 1300
},
{
"decision_loss": 4.3113,
"encoder_mlm_loss": 0.7483,
"epoch": 0.07,
"grad_norm": 13.525325775146484,
"learning_rate": 1.341e-05,
"loss": 11.2692,
"reason_loss": NaN,
"step": 1350
},
{
"decision_loss": 4.4514,
"encoder_mlm_loss": 0.7134,
"epoch": 0.07,
"grad_norm": 12.281783103942871,
"learning_rate": 1.391e-05,
"loss": 11.3637,
"reason_loss": NaN,
"step": 1400
},
{
"decision_loss": 4.2842,
"encoder_mlm_loss": 0.707,
"epoch": 0.07,
"grad_norm": 13.143014907836914,
"learning_rate": 1.4410000000000001e-05,
"loss": 11.083,
"reason_loss": NaN,
"step": 1450
},
{
"decision_loss": 4.1891,
"encoder_mlm_loss": 0.7466,
"epoch": 0.08,
"grad_norm": 15.713329315185547,
"learning_rate": 1.4910000000000001e-05,
"loss": 11.0169,
"reason_loss": NaN,
"step": 1500
},
{
"decision_loss": 4.2626,
"encoder_mlm_loss": 0.7105,
"epoch": 0.08,
"grad_norm": 12.946088790893555,
"learning_rate": 1.541e-05,
"loss": 10.9707,
"reason_loss": NaN,
"step": 1550
},
{
"decision_loss": 4.2396,
"encoder_mlm_loss": 0.7035,
"epoch": 0.08,
"grad_norm": 14.146540641784668,
"learning_rate": 1.591e-05,
"loss": 10.8574,
"reason_loss": NaN,
"step": 1600
},
{
"decision_loss": 4.2699,
"encoder_mlm_loss": 0.6882,
"epoch": 0.09,
"grad_norm": 13.31047534942627,
"learning_rate": 1.641e-05,
"loss": 11.0437,
"reason_loss": NaN,
"step": 1650
},
{
"decision_loss": 4.2453,
"encoder_mlm_loss": 0.6995,
"epoch": 0.09,
"grad_norm": 12.856541633605957,
"learning_rate": 1.6910000000000002e-05,
"loss": 10.9505,
"reason_loss": NaN,
"step": 1700
},
{
"decision_loss": 4.3523,
"encoder_mlm_loss": 0.6896,
"epoch": 0.09,
"grad_norm": 13.268596649169922,
"learning_rate": 1.741e-05,
"loss": 10.9844,
"reason_loss": NaN,
"step": 1750
},
{
"decision_loss": 4.262,
"encoder_mlm_loss": 0.6877,
"epoch": 0.09,
"grad_norm": 13.404997825622559,
"learning_rate": 1.7910000000000003e-05,
"loss": 10.9554,
"reason_loss": NaN,
"step": 1800
},
{
"decision_loss": 4.2655,
"encoder_mlm_loss": 0.6793,
"epoch": 0.1,
"grad_norm": 12.608678817749023,
"learning_rate": 1.841e-05,
"loss": 10.9086,
"reason_loss": NaN,
"step": 1850
},
{
"decision_loss": 4.2445,
"encoder_mlm_loss": 0.6692,
"epoch": 0.1,
"grad_norm": 11.864866256713867,
"learning_rate": 1.891e-05,
"loss": 11.0079,
"reason_loss": NaN,
"step": 1900
},
{
"decision_loss": 4.2461,
"encoder_mlm_loss": 0.7055,
"epoch": 0.1,
"grad_norm": 12.110335350036621,
"learning_rate": 1.941e-05,
"loss": 10.9009,
"reason_loss": NaN,
"step": 1950
},
{
"decision_loss": 4.1508,
"encoder_mlm_loss": 0.6584,
"epoch": 0.1,
"grad_norm": 13.818020820617676,
"learning_rate": 1.991e-05,
"loss": 10.9391,
"reason_loss": NaN,
"step": 2000
},
{
"decision_loss": 4.1503,
"encoder_mlm_loss": 0.6471,
"epoch": 0.11,
"grad_norm": 11.85276985168457,
"learning_rate": 2.0410000000000003e-05,
"loss": 10.703,
"reason_loss": NaN,
"step": 2050
},
{
"decision_loss": 4.1422,
"encoder_mlm_loss": 0.6624,
"epoch": 0.11,
"grad_norm": 11.124549865722656,
"learning_rate": 2.091e-05,
"loss": 10.7727,
"reason_loss": NaN,
"step": 2100
},
{
"decision_loss": 4.1424,
"encoder_mlm_loss": 0.6796,
"epoch": 0.11,
"grad_norm": 12.21742057800293,
"learning_rate": 2.1410000000000003e-05,
"loss": 10.6846,
"reason_loss": NaN,
"step": 2150
},
{
"decision_loss": 4.0898,
"encoder_mlm_loss": 0.6717,
"epoch": 0.11,
"grad_norm": 11.365450859069824,
"learning_rate": 2.191e-05,
"loss": 10.662,
"reason_loss": NaN,
"step": 2200
},
{
"decision_loss": 4.0713,
"encoder_mlm_loss": 0.6881,
"epoch": 0.12,
"grad_norm": 12.824188232421875,
"learning_rate": 2.241e-05,
"loss": 10.8777,
"reason_loss": NaN,
"step": 2250
},
{
"decision_loss": 4.1075,
"encoder_mlm_loss": 0.6395,
"epoch": 0.12,
"grad_norm": 12.549530982971191,
"learning_rate": 2.2910000000000003e-05,
"loss": 10.6139,
"reason_loss": NaN,
"step": 2300
},
{
"decision_loss": 4.1325,
"encoder_mlm_loss": 0.6705,
"epoch": 0.12,
"grad_norm": 11.592597007751465,
"learning_rate": 2.341e-05,
"loss": 10.6118,
"reason_loss": NaN,
"step": 2350
},
{
"decision_loss": 4.0534,
"encoder_mlm_loss": 0.6424,
"epoch": 0.12,
"grad_norm": 12.411306381225586,
"learning_rate": 2.3910000000000003e-05,
"loss": 10.586,
"reason_loss": NaN,
"step": 2400
},
{
"decision_loss": 4.0439,
"encoder_mlm_loss": 0.6486,
"epoch": 0.13,
"grad_norm": 12.624602317810059,
"learning_rate": 2.4410000000000002e-05,
"loss": 10.5166,
"reason_loss": NaN,
"step": 2450
},
{
"decision_loss": 4.1069,
"encoder_mlm_loss": 0.6389,
"epoch": 0.13,
"grad_norm": 13.636293411254883,
"learning_rate": 2.491e-05,
"loss": 10.6555,
"reason_loss": NaN,
"step": 2500
},
{
"decision_loss": 4.0365,
"encoder_mlm_loss": 0.6552,
"epoch": 0.13,
"grad_norm": 11.545199394226074,
"learning_rate": 2.541e-05,
"loss": 10.4713,
"reason_loss": NaN,
"step": 2550
},
{
"decision_loss": 4.0198,
"encoder_mlm_loss": 0.6485,
"epoch": 0.13,
"grad_norm": 11.711776733398438,
"learning_rate": 2.591e-05,
"loss": 10.5863,
"reason_loss": NaN,
"step": 2600
},
{
"decision_loss": 3.9609,
"encoder_mlm_loss": 0.6039,
"epoch": 0.14,
"grad_norm": 13.17541217803955,
"learning_rate": 2.6410000000000003e-05,
"loss": 10.3168,
"reason_loss": NaN,
"step": 2650
},
{
"decision_loss": 4.0235,
"encoder_mlm_loss": 0.6256,
"epoch": 0.14,
"grad_norm": 11.82180404663086,
"learning_rate": 2.691e-05,
"loss": 10.5624,
"reason_loss": NaN,
"step": 2700
},
{
"decision_loss": 3.839,
"encoder_mlm_loss": 0.6519,
"epoch": 0.14,
"grad_norm": 12.721589088439941,
"learning_rate": 2.7410000000000004e-05,
"loss": 10.4343,
"reason_loss": NaN,
"step": 2750
},
{
"decision_loss": 3.9386,
"encoder_mlm_loss": 0.6115,
"epoch": 0.14,
"grad_norm": 12.1082763671875,
"learning_rate": 2.7910000000000002e-05,
"loss": 10.3518,
"reason_loss": NaN,
"step": 2800
},
{
"decision_loss": 3.8515,
"encoder_mlm_loss": 0.6094,
"epoch": 0.15,
"grad_norm": 12.88183307647705,
"learning_rate": 2.8410000000000004e-05,
"loss": 10.2022,
"reason_loss": NaN,
"step": 2850
},
{
"decision_loss": 3.7709,
"encoder_mlm_loss": 0.6229,
"epoch": 0.15,
"grad_norm": 11.533916473388672,
"learning_rate": 2.8899999999999998e-05,
"loss": 10.3799,
"reason_loss": NaN,
"step": 2900
},
{
"decision_loss": 3.8497,
"encoder_mlm_loss": 0.6279,
"epoch": 0.15,
"grad_norm": 12.61514663696289,
"learning_rate": 2.94e-05,
"loss": 10.1719,
"reason_loss": NaN,
"step": 2950
},
{
"decision_loss": 3.8989,
"encoder_mlm_loss": 0.6137,
"epoch": 0.15,
"grad_norm": 10.787571907043457,
"learning_rate": 2.9900000000000002e-05,
"loss": 10.3554,
"reason_loss": NaN,
"step": 3000
},
{
"decision_loss": 3.7403,
"encoder_mlm_loss": 0.6325,
"epoch": 0.16,
"grad_norm": 11.576037406921387,
"learning_rate": 3.04e-05,
"loss": 10.1061,
"reason_loss": NaN,
"step": 3050
},
{
"decision_loss": 3.8257,
"encoder_mlm_loss": 0.6188,
"epoch": 0.16,
"grad_norm": 11.931550025939941,
"learning_rate": 3.09e-05,
"loss": 10.0557,
"reason_loss": NaN,
"step": 3100
},
{
"decision_loss": 3.7653,
"encoder_mlm_loss": 0.6151,
"epoch": 0.16,
"grad_norm": 13.052062034606934,
"learning_rate": 3.1400000000000004e-05,
"loss": 9.9838,
"reason_loss": NaN,
"step": 3150
},
{
"decision_loss": 3.7544,
"encoder_mlm_loss": 0.6237,
"epoch": 0.16,
"grad_norm": 14.21654224395752,
"learning_rate": 3.19e-05,
"loss": 9.8657,
"reason_loss": NaN,
"step": 3200
},
{
"decision_loss": 3.6772,
"encoder_mlm_loss": 0.61,
"epoch": 0.17,
"grad_norm": 14.428752899169922,
"learning_rate": 3.24e-05,
"loss": 9.6336,
"reason_loss": NaN,
"step": 3250
},
{
"decision_loss": 3.6212,
"encoder_mlm_loss": 0.6203,
"epoch": 0.17,
"grad_norm": 15.353618621826172,
"learning_rate": 3.29e-05,
"loss": 9.4442,
"reason_loss": NaN,
"step": 3300
},
{
"decision_loss": 3.6407,
"encoder_mlm_loss": 0.6454,
"epoch": 0.17,
"grad_norm": 12.828969955444336,
"learning_rate": 3.3400000000000005e-05,
"loss": 9.3448,
"reason_loss": NaN,
"step": 3350
},
{
"decision_loss": 3.5022,
"encoder_mlm_loss": 0.6361,
"epoch": 0.18,
"grad_norm": 13.936616897583008,
"learning_rate": 3.3900000000000004e-05,
"loss": 9.354,
"reason_loss": NaN,
"step": 3400
},
{
"decision_loss": 3.5961,
"encoder_mlm_loss": 0.6405,
"epoch": 0.18,
"grad_norm": 13.668755531311035,
"learning_rate": 3.4399999999999996e-05,
"loss": 9.0563,
"reason_loss": 4.3796,
"step": 3450
},
{
"decision_loss": 3.5257,
"encoder_mlm_loss": 0.591,
"epoch": 0.18,
"grad_norm": 12.743650436401367,
"learning_rate": 3.49e-05,
"loss": 9.089,
"reason_loss": NaN,
"step": 3500
},
{
"decision_loss": 3.688,
"encoder_mlm_loss": 0.6529,
"epoch": 0.18,
"grad_norm": 13.899826049804688,
"learning_rate": 3.54e-05,
"loss": 8.7378,
"reason_loss": NaN,
"step": 3550
},
{
"decision_loss": 3.5177,
"encoder_mlm_loss": 0.6223,
"epoch": 0.19,
"grad_norm": 16.01102066040039,
"learning_rate": 3.59e-05,
"loss": 8.6978,
"reason_loss": NaN,
"step": 3600
},
{
"decision_loss": 3.5152,
"encoder_mlm_loss": 0.6192,
"epoch": 0.19,
"grad_norm": 14.593777656555176,
"learning_rate": 3.6400000000000004e-05,
"loss": 8.3674,
"reason_loss": NaN,
"step": 3650
},
{
"decision_loss": 3.4812,
"encoder_mlm_loss": 0.6087,
"epoch": 0.19,
"grad_norm": 20.128116607666016,
"learning_rate": 3.69e-05,
"loss": 8.2724,
"reason_loss": NaN,
"step": 3700
},
{
"decision_loss": 3.5961,
"encoder_mlm_loss": 0.596,
"epoch": 0.19,
"grad_norm": 15.367920875549316,
"learning_rate": 3.74e-05,
"loss": 8.1541,
"reason_loss": NaN,
"step": 3750
},
{
"decision_loss": 3.438,
"encoder_mlm_loss": 0.6147,
"epoch": 0.2,
"grad_norm": 13.092328071594238,
"learning_rate": 3.79e-05,
"loss": 8.0694,
"reason_loss": NaN,
"step": 3800
},
{
"decision_loss": 3.3987,
"encoder_mlm_loss": 0.6024,
"epoch": 0.2,
"grad_norm": 14.642455101013184,
"learning_rate": 3.8400000000000005e-05,
"loss": 7.8633,
"reason_loss": NaN,
"step": 3850
},
{
"decision_loss": 3.3564,
"encoder_mlm_loss": 0.5611,
"epoch": 0.2,
"grad_norm": 14.328669548034668,
"learning_rate": 3.8900000000000004e-05,
"loss": 7.4975,
"reason_loss": NaN,
"step": 3900
},
{
"decision_loss": 3.4611,
"encoder_mlm_loss": 0.6012,
"epoch": 0.2,
"grad_norm": 13.032312393188477,
"learning_rate": 3.94e-05,
"loss": 7.4354,
"reason_loss": NaN,
"step": 3950
},
{
"decision_loss": 3.3266,
"encoder_mlm_loss": 0.6198,
"epoch": 0.21,
"grad_norm": 13.335681915283203,
"learning_rate": 3.99e-05,
"loss": 7.3149,
"reason_loss": NaN,
"step": 4000
},
{
"decision_loss": 3.3579,
"encoder_mlm_loss": 0.589,
"epoch": 0.21,
"grad_norm": 13.334540367126465,
"learning_rate": 4.0400000000000006e-05,
"loss": 7.2804,
"reason_loss": NaN,
"step": 4050
},
{
"decision_loss": 3.4332,
"encoder_mlm_loss": 0.5642,
"epoch": 0.21,
"grad_norm": 13.725900650024414,
"learning_rate": 4.09e-05,
"loss": 7.2422,
"reason_loss": NaN,
"step": 4100
},
{
"decision_loss": 3.3579,
"encoder_mlm_loss": 0.638,
"epoch": 0.21,
"grad_norm": 12.601729393005371,
"learning_rate": 4.14e-05,
"loss": 7.2021,
"reason_loss": NaN,
"step": 4150
},
{
"decision_loss": 3.3385,
"encoder_mlm_loss": 0.6249,
"epoch": 0.22,
"grad_norm": 14.782654762268066,
"learning_rate": 4.19e-05,
"loss": 7.1519,
"reason_loss": NaN,
"step": 4200
},
{
"decision_loss": 3.2423,
"encoder_mlm_loss": 0.572,
"epoch": 0.22,
"grad_norm": 10.786784172058105,
"learning_rate": 4.24e-05,
"loss": 7.0259,
"reason_loss": NaN,
"step": 4250
},
{
"decision_loss": 3.4216,
"encoder_mlm_loss": 0.6259,
"epoch": 0.22,
"grad_norm": 14.012487411499023,
"learning_rate": 4.29e-05,
"loss": 6.8396,
"reason_loss": NaN,
"step": 4300
},
{
"decision_loss": 3.294,
"encoder_mlm_loss": 0.6021,
"epoch": 0.22,
"grad_norm": 13.132121086120605,
"learning_rate": 4.3400000000000005e-05,
"loss": 6.7559,
"reason_loss": NaN,
"step": 4350
},
{
"decision_loss": 3.2271,
"encoder_mlm_loss": 0.6381,
"epoch": 0.23,
"grad_norm": 14.553518295288086,
"learning_rate": 4.39e-05,
"loss": 6.8517,
"reason_loss": 2.5116,
"step": 4400
},
{
"decision_loss": 3.1722,
"encoder_mlm_loss": 0.5848,
"epoch": 0.23,
"grad_norm": 13.60247802734375,
"learning_rate": 4.44e-05,
"loss": 6.6558,
"reason_loss": NaN,
"step": 4450
},
{
"decision_loss": 3.3701,
"encoder_mlm_loss": 0.6151,
"epoch": 0.23,
"grad_norm": 14.23380184173584,
"learning_rate": 4.49e-05,
"loss": 6.6485,
"reason_loss": NaN,
"step": 4500
},
{
"decision_loss": 3.214,
"encoder_mlm_loss": 0.5903,
"epoch": 0.23,
"grad_norm": 12.637818336486816,
"learning_rate": 4.5400000000000006e-05,
"loss": 6.6434,
"reason_loss": NaN,
"step": 4550
},
{
"decision_loss": 3.4377,
"encoder_mlm_loss": 0.5688,
"epoch": 0.24,
"grad_norm": 12.161514282226562,
"learning_rate": 4.5900000000000004e-05,
"loss": 6.558,
"reason_loss": NaN,
"step": 4600
},
{
"decision_loss": 3.2337,
"encoder_mlm_loss": 0.5745,
"epoch": 0.24,
"grad_norm": 10.774749755859375,
"learning_rate": 4.64e-05,
"loss": 6.4569,
"reason_loss": NaN,
"step": 4650
},
{
"decision_loss": 3.1934,
"encoder_mlm_loss": 0.6216,
"epoch": 0.24,
"grad_norm": 12.889678001403809,
"learning_rate": 4.69e-05,
"loss": 6.588,
"reason_loss": NaN,
"step": 4700
},
{
"decision_loss": 3.2473,
"encoder_mlm_loss": 0.598,
"epoch": 0.24,
"grad_norm": 12.259153366088867,
"learning_rate": 4.74e-05,
"loss": 6.506,
"reason_loss": NaN,
"step": 4750
},
{
"decision_loss": 3.3514,
"encoder_mlm_loss": 0.5984,
"epoch": 0.25,
"grad_norm": 13.196783065795898,
"learning_rate": 4.79e-05,
"loss": 6.4249,
"reason_loss": NaN,
"step": 4800
},
{
"decision_loss": 3.125,
"encoder_mlm_loss": 0.5777,
"epoch": 0.25,
"grad_norm": 12.129148483276367,
"learning_rate": 4.8400000000000004e-05,
"loss": 6.1404,
"reason_loss": NaN,
"step": 4850
},
{
"decision_loss": 3.184,
"encoder_mlm_loss": 0.5889,
"epoch": 0.25,
"grad_norm": 12.139684677124023,
"learning_rate": 4.89e-05,
"loss": 6.1487,
"reason_loss": NaN,
"step": 4900
},
{
"decision_loss": 3.2688,
"encoder_mlm_loss": 0.583,
"epoch": 0.26,
"grad_norm": 12.38558578491211,
"learning_rate": 4.94e-05,
"loss": 6.2462,
"reason_loss": NaN,
"step": 4950
},
{
"decision_loss": 3.0885,
"encoder_mlm_loss": 0.5828,
"epoch": 0.26,
"grad_norm": 12.759172439575195,
"learning_rate": 4.99e-05,
"loss": 6.2834,
"reason_loss": NaN,
"step": 5000
},
{
"decision_loss": 3.1414,
"encoder_mlm_loss": 0.5995,
"epoch": 0.26,
"grad_norm": 13.554661750793457,
"learning_rate": 4.986107252014449e-05,
"loss": 6.158,
"reason_loss": NaN,
"step": 5050
},
{
"decision_loss": 3.0832,
"encoder_mlm_loss": 0.5885,
"epoch": 0.26,
"grad_norm": 14.811614990234375,
"learning_rate": 4.9687413170325095e-05,
"loss": 6.0866,
"reason_loss": NaN,
"step": 5100
},
{
"decision_loss": 3.0845,
"encoder_mlm_loss": 0.5737,
"epoch": 0.27,
"grad_norm": 11.42420482635498,
"learning_rate": 4.9513753820505696e-05,
"loss": 6.0264,
"reason_loss": NaN,
"step": 5150
},
{
"decision_loss": 3.0805,
"encoder_mlm_loss": 0.583,
"epoch": 0.27,
"grad_norm": 12.164910316467285,
"learning_rate": 4.9340094470686304e-05,
"loss": 6.0071,
"reason_loss": NaN,
"step": 5200
},
{
"decision_loss": 3.0178,
"encoder_mlm_loss": 0.5798,
"epoch": 0.27,
"grad_norm": 12.894425392150879,
"learning_rate": 4.916643512086691e-05,
"loss": 5.9767,
"reason_loss": NaN,
"step": 5250
},
{
"decision_loss": 2.9613,
"encoder_mlm_loss": 0.5629,
"epoch": 0.27,
"grad_norm": 12.356279373168945,
"learning_rate": 4.899277577104751e-05,
"loss": 5.9532,
"reason_loss": NaN,
"step": 5300
},
{
"decision_loss": 2.9668,
"encoder_mlm_loss": 0.5662,
"epoch": 0.28,
"grad_norm": 12.371698379516602,
"learning_rate": 4.881911642122812e-05,
"loss": 5.9129,
"reason_loss": NaN,
"step": 5350
},
{
"decision_loss": 2.8377,
"encoder_mlm_loss": 0.5764,
"epoch": 0.28,
"grad_norm": 11.876270294189453,
"learning_rate": 4.864545707140873e-05,
"loss": 5.7633,
"reason_loss": NaN,
"step": 5400
},
{
"decision_loss": 3.0811,
"encoder_mlm_loss": 0.5976,
"epoch": 0.28,
"grad_norm": 11.529359817504883,
"learning_rate": 4.847179772158934e-05,
"loss": 5.9185,
"reason_loss": NaN,
"step": 5450
},
{
"decision_loss": 2.931,
"encoder_mlm_loss": 0.5827,
"epoch": 0.28,
"grad_norm": 12.832273483276367,
"learning_rate": 4.829813837176994e-05,
"loss": 5.9475,
"reason_loss": NaN,
"step": 5500
},
{
"decision_loss": 2.9643,
"encoder_mlm_loss": 0.5972,
"epoch": 0.29,
"grad_norm": 12.522998809814453,
"learning_rate": 4.812447902195054e-05,
"loss": 5.9335,
"reason_loss": NaN,
"step": 5550
},
{
"decision_loss": 2.9406,
"encoder_mlm_loss": 0.5756,
"epoch": 0.29,
"grad_norm": 11.039345741271973,
"learning_rate": 4.795081967213115e-05,
"loss": 5.7175,
"reason_loss": NaN,
"step": 5600
},
{
"decision_loss": 2.8629,
"encoder_mlm_loss": 0.5527,
"epoch": 0.29,
"grad_norm": 12.582688331604004,
"learning_rate": 4.7777160322311755e-05,
"loss": 5.6501,
"reason_loss": NaN,
"step": 5650
},
{
"decision_loss": 2.8334,
"encoder_mlm_loss": 0.5859,
"epoch": 0.29,
"grad_norm": 12.082906723022461,
"learning_rate": 4.760350097249236e-05,
"loss": 5.7244,
"reason_loss": NaN,
"step": 5700
},
{
"decision_loss": 2.8741,
"encoder_mlm_loss": 0.5826,
"epoch": 0.3,
"grad_norm": 11.20120620727539,
"learning_rate": 4.7429841622672964e-05,
"loss": 5.6161,
"reason_loss": NaN,
"step": 5750
},
{
"decision_loss": 2.9188,
"encoder_mlm_loss": 0.5634,
"epoch": 0.3,
"grad_norm": 13.118001937866211,
"learning_rate": 4.725618227285357e-05,
"loss": 5.6669,
"reason_loss": NaN,
"step": 5800
},
{
"decision_loss": 2.9045,
"encoder_mlm_loss": 0.5613,
"epoch": 0.3,
"grad_norm": 11.6439208984375,
"learning_rate": 4.708252292303418e-05,
"loss": 5.6609,
"reason_loss": NaN,
"step": 5850
},
{
"decision_loss": 2.9099,
"encoder_mlm_loss": 0.5681,
"epoch": 0.3,
"grad_norm": 11.001469612121582,
"learning_rate": 4.690886357321478e-05,
"loss": 5.5608,
"reason_loss": NaN,
"step": 5900
},
{
"decision_loss": 2.8081,
"encoder_mlm_loss": 0.5748,
"epoch": 0.31,
"grad_norm": 11.262368202209473,
"learning_rate": 4.673520422339539e-05,
"loss": 5.6866,
"reason_loss": NaN,
"step": 5950
},
{
"decision_loss": 2.9234,
"encoder_mlm_loss": 0.5752,
"epoch": 0.31,
"grad_norm": 15.921095848083496,
"learning_rate": 4.6561544873576e-05,
"loss": 5.4638,
"reason_loss": NaN,
"step": 6000
},
{
"decision_loss": 2.8734,
"encoder_mlm_loss": 0.5436,
"epoch": 0.31,
"grad_norm": 11.366347312927246,
"learning_rate": 4.6387885523756605e-05,
"loss": 5.4932,
"reason_loss": NaN,
"step": 6050
},
{
"decision_loss": 2.8849,
"encoder_mlm_loss": 0.5758,
"epoch": 0.31,
"grad_norm": 13.834676742553711,
"learning_rate": 4.6214226173937206e-05,
"loss": 5.5872,
"reason_loss": NaN,
"step": 6100
},
{
"decision_loss": 2.8112,
"encoder_mlm_loss": 0.5829,
"epoch": 0.32,
"grad_norm": 12.526769638061523,
"learning_rate": 4.6040566824117814e-05,
"loss": 5.4441,
"reason_loss": NaN,
"step": 6150
},
{
"decision_loss": 2.7713,
"encoder_mlm_loss": 0.5631,
"epoch": 0.32,
"grad_norm": 11.834070205688477,
"learning_rate": 4.586690747429842e-05,
"loss": 5.4597,
"reason_loss": NaN,
"step": 6200
},
{
"decision_loss": 2.8737,
"encoder_mlm_loss": 0.5454,
"epoch": 0.32,
"grad_norm": 10.592991828918457,
"learning_rate": 4.569324812447903e-05,
"loss": 5.4016,
"reason_loss": NaN,
"step": 6250
},
{
"decision_loss": 2.7021,
"encoder_mlm_loss": 0.5667,
"epoch": 0.32,
"grad_norm": 11.339573860168457,
"learning_rate": 4.5519588774659624e-05,
"loss": 5.5297,
"reason_loss": NaN,
"step": 6300
},
{
"decision_loss": 2.8387,
"encoder_mlm_loss": 0.5498,
"epoch": 0.33,
"grad_norm": 14.197671890258789,
"learning_rate": 4.534592942484023e-05,
"loss": 5.3895,
"reason_loss": 1.7433,
"step": 6350
},
{
"decision_loss": 2.6152,
"encoder_mlm_loss": 0.555,
"epoch": 0.33,
"grad_norm": 12.190983772277832,
"learning_rate": 4.517227007502084e-05,
"loss": 5.342,
"reason_loss": NaN,
"step": 6400
},
{
"decision_loss": 2.8126,
"encoder_mlm_loss": 0.5843,
"epoch": 0.33,
"grad_norm": 20.80158805847168,
"learning_rate": 4.499861072520145e-05,
"loss": 5.4744,
"reason_loss": NaN,
"step": 6450
},
{
"decision_loss": 2.8011,
"encoder_mlm_loss": 0.5559,
"epoch": 0.34,
"grad_norm": 10.541352272033691,
"learning_rate": 4.482495137538205e-05,
"loss": 5.3561,
"reason_loss": NaN,
"step": 6500
},
{
"decision_loss": 2.7874,
"encoder_mlm_loss": 0.5694,
"epoch": 0.34,
"grad_norm": 15.321596145629883,
"learning_rate": 4.465129202556266e-05,
"loss": 5.3279,
"reason_loss": NaN,
"step": 6550
},
{
"decision_loss": 2.7497,
"encoder_mlm_loss": 0.5539,
"epoch": 0.34,
"grad_norm": 12.009075164794922,
"learning_rate": 4.4477632675743265e-05,
"loss": 5.3333,
"reason_loss": NaN,
"step": 6600
},
{
"decision_loss": 2.641,
"encoder_mlm_loss": 0.5661,
"epoch": 0.34,
"grad_norm": 19.22854995727539,
"learning_rate": 4.430397332592387e-05,
"loss": 5.3703,
"reason_loss": NaN,
"step": 6650
},
{
"decision_loss": 2.6307,
"encoder_mlm_loss": 0.541,
"epoch": 0.35,
"grad_norm": 11.551212310791016,
"learning_rate": 4.4130313976104474e-05,
"loss": 5.3032,
"reason_loss": NaN,
"step": 6700
},
{
"decision_loss": 2.852,
"encoder_mlm_loss": 0.5668,
"epoch": 0.35,
"grad_norm": 11.5441312789917,
"learning_rate": 4.395665462628508e-05,
"loss": 5.38,
"reason_loss": NaN,
"step": 6750
},
{
"decision_loss": 2.7921,
"encoder_mlm_loss": 0.5515,
"epoch": 0.35,
"grad_norm": 10.488402366638184,
"learning_rate": 4.378299527646569e-05,
"loss": 5.211,
"reason_loss": NaN,
"step": 6800
},
{
"decision_loss": 2.7793,
"encoder_mlm_loss": 0.5353,
"epoch": 0.35,
"grad_norm": 11.62839412689209,
"learning_rate": 4.36093359266463e-05,
"loss": 5.2583,
"reason_loss": NaN,
"step": 6850
},
{
"decision_loss": 2.5967,
"encoder_mlm_loss": 0.5202,
"epoch": 0.36,
"grad_norm": 11.641283988952637,
"learning_rate": 4.34356765768269e-05,
"loss": 5.1396,
"reason_loss": NaN,
"step": 6900
},
{
"decision_loss": 2.6503,
"encoder_mlm_loss": 0.5485,
"epoch": 0.36,
"grad_norm": 12.255797386169434,
"learning_rate": 4.3262017227007506e-05,
"loss": 5.2417,
"reason_loss": NaN,
"step": 6950
},
{
"decision_loss": 2.6204,
"encoder_mlm_loss": 0.5421,
"epoch": 0.36,
"grad_norm": 10.042210578918457,
"learning_rate": 4.3088357877188114e-05,
"loss": 5.1232,
"reason_loss": NaN,
"step": 7000
},
{
"decision_loss": 2.7291,
"encoder_mlm_loss": 0.5607,
"epoch": 0.36,
"grad_norm": 10.926862716674805,
"learning_rate": 4.2914698527368716e-05,
"loss": 5.1705,
"reason_loss": NaN,
"step": 7050
},
{
"decision_loss": 2.7328,
"encoder_mlm_loss": 0.5431,
"epoch": 0.37,
"grad_norm": 11.532885551452637,
"learning_rate": 4.274103917754932e-05,
"loss": 5.0906,
"reason_loss": NaN,
"step": 7100
},
{
"decision_loss": 2.5302,
"encoder_mlm_loss": 0.531,
"epoch": 0.37,
"grad_norm": 11.01016902923584,
"learning_rate": 4.2567379827729925e-05,
"loss": 4.9902,
"reason_loss": NaN,
"step": 7150
},
{
"decision_loss": 2.7506,
"encoder_mlm_loss": 0.5675,
"epoch": 0.37,
"grad_norm": 11.220144271850586,
"learning_rate": 4.239372047791053e-05,
"loss": 5.2234,
"reason_loss": NaN,
"step": 7200
},
{
"decision_loss": 2.6282,
"encoder_mlm_loss": 0.5864,
"epoch": 0.37,
"grad_norm": 10.498862266540527,
"learning_rate": 4.222006112809114e-05,
"loss": 5.1675,
"reason_loss": NaN,
"step": 7250
},
{
"decision_loss": 2.5942,
"encoder_mlm_loss": 0.5078,
"epoch": 0.38,
"grad_norm": 11.325239181518555,
"learning_rate": 4.204640177827174e-05,
"loss": 5.093,
"reason_loss": NaN,
"step": 7300
},
{
"decision_loss": 2.6822,
"encoder_mlm_loss": 0.509,
"epoch": 0.38,
"grad_norm": 10.894235610961914,
"learning_rate": 4.187274242845235e-05,
"loss": 5.1333,
"reason_loss": NaN,
"step": 7350
},
{
"decision_loss": 2.6609,
"encoder_mlm_loss": 0.5532,
"epoch": 0.38,
"grad_norm": 10.849638938903809,
"learning_rate": 4.169908307863296e-05,
"loss": 5.0219,
"reason_loss": NaN,
"step": 7400
},
{
"decision_loss": 2.7642,
"encoder_mlm_loss": 0.5628,
"epoch": 0.38,
"grad_norm": 9.63953685760498,
"learning_rate": 4.152542372881356e-05,
"loss": 5.0,
"reason_loss": NaN,
"step": 7450
},
{
"decision_loss": 2.629,
"encoder_mlm_loss": 0.5054,
"epoch": 0.39,
"grad_norm": 12.991473197937012,
"learning_rate": 4.1351764378994166e-05,
"loss": 5.0013,
"reason_loss": NaN,
"step": 7500
},
{
"decision_loss": 2.6034,
"encoder_mlm_loss": 0.5576,
"epoch": 0.39,
"grad_norm": 11.331198692321777,
"learning_rate": 4.1178105029174774e-05,
"loss": 5.1522,
"reason_loss": NaN,
"step": 7550
},
{
"decision_loss": 2.5915,
"encoder_mlm_loss": 0.5671,
"epoch": 0.39,
"grad_norm": 10.665949821472168,
"learning_rate": 4.100444567935538e-05,
"loss": 4.9226,
"reason_loss": NaN,
"step": 7600
},
{
"decision_loss": 2.6619,
"encoder_mlm_loss": 0.5314,
"epoch": 0.39,
"grad_norm": 9.840173721313477,
"learning_rate": 4.083078632953598e-05,
"loss": 4.98,
"reason_loss": NaN,
"step": 7650
},
{
"decision_loss": 2.7066,
"encoder_mlm_loss": 0.5427,
"epoch": 0.4,
"grad_norm": 10.543719291687012,
"learning_rate": 4.065712697971659e-05,
"loss": 4.9124,
"reason_loss": NaN,
"step": 7700
},
{
"decision_loss": 2.5665,
"encoder_mlm_loss": 0.5278,
"epoch": 0.4,
"grad_norm": 11.529766082763672,
"learning_rate": 4.04834676298972e-05,
"loss": 4.9287,
"reason_loss": NaN,
"step": 7750
},
{
"decision_loss": 2.5556,
"encoder_mlm_loss": 0.5677,
"epoch": 0.4,
"grad_norm": 11.010191917419434,
"learning_rate": 4.03098082800778e-05,
"loss": 4.9821,
"reason_loss": NaN,
"step": 7800
},
{
"decision_loss": 2.5638,
"encoder_mlm_loss": 0.5572,
"epoch": 0.4,
"grad_norm": 11.854859352111816,
"learning_rate": 4.013614893025841e-05,
"loss": 4.8495,
"reason_loss": NaN,
"step": 7850
},
{
"decision_loss": 2.5787,
"encoder_mlm_loss": 0.524,
"epoch": 0.41,
"grad_norm": 15.032380104064941,
"learning_rate": 3.996248958043901e-05,
"loss": 4.9092,
"reason_loss": NaN,
"step": 7900
},
{
"decision_loss": 2.6208,
"encoder_mlm_loss": 0.5749,
"epoch": 0.41,
"grad_norm": 10.260398864746094,
"learning_rate": 3.978883023061962e-05,
"loss": 4.96,
"reason_loss": NaN,
"step": 7950
},
{
"decision_loss": 2.5406,
"encoder_mlm_loss": 0.5159,
"epoch": 0.41,
"grad_norm": 10.527961730957031,
"learning_rate": 3.9615170880800225e-05,
"loss": 5.0116,
"reason_loss": NaN,
"step": 8000
},
{
"decision_loss": 2.5387,
"encoder_mlm_loss": 0.56,
"epoch": 0.42,
"grad_norm": 10.906159400939941,
"learning_rate": 3.9441511530980826e-05,
"loss": 4.9037,
"reason_loss": NaN,
"step": 8050
},
{
"decision_loss": 2.5531,
"encoder_mlm_loss": 0.5345,
"epoch": 0.42,
"grad_norm": 11.020978927612305,
"learning_rate": 3.9267852181161434e-05,
"loss": 4.7901,
"reason_loss": NaN,
"step": 8100
},
{
"decision_loss": 2.3603,
"encoder_mlm_loss": 0.5403,
"epoch": 0.42,
"grad_norm": 9.852022171020508,
"learning_rate": 3.909419283134204e-05,
"loss": 4.8887,
"reason_loss": NaN,
"step": 8150
},
{
"decision_loss": 2.5552,
"encoder_mlm_loss": 0.5263,
"epoch": 0.42,
"grad_norm": 10.380234718322754,
"learning_rate": 3.892053348152265e-05,
"loss": 4.823,
"reason_loss": NaN,
"step": 8200
},
{
"decision_loss": 2.483,
"encoder_mlm_loss": 0.5314,
"epoch": 0.43,
"grad_norm": 15.00808334350586,
"learning_rate": 3.874687413170325e-05,
"loss": 4.8393,
"reason_loss": NaN,
"step": 8250
},
{
"decision_loss": 2.5897,
"encoder_mlm_loss": 0.5297,
"epoch": 0.43,
"grad_norm": 11.162062644958496,
"learning_rate": 3.857321478188386e-05,
"loss": 4.8635,
"reason_loss": NaN,
"step": 8300
},
{
"decision_loss": 2.5224,
"encoder_mlm_loss": 0.5486,
"epoch": 0.43,
"grad_norm": 10.835860252380371,
"learning_rate": 3.8403028619060855e-05,
"loss": 4.7197,
"reason_loss": NaN,
"step": 8350
},
{
"decision_loss": 2.5729,
"encoder_mlm_loss": 0.5287,
"epoch": 0.43,
"grad_norm": 10.745417594909668,
"learning_rate": 3.8229369269241456e-05,
"loss": 4.863,
"reason_loss": NaN,
"step": 8400
},
{
"decision_loss": 2.5483,
"encoder_mlm_loss": 0.4999,
"epoch": 0.44,
"grad_norm": 11.162773132324219,
"learning_rate": 3.8055709919422064e-05,
"loss": 4.7945,
"reason_loss": NaN,
"step": 8450
},
{
"decision_loss": 2.4687,
"encoder_mlm_loss": 0.5278,
"epoch": 0.44,
"grad_norm": 9.60230541229248,
"learning_rate": 3.7882050569602665e-05,
"loss": 4.9069,
"reason_loss": NaN,
"step": 8500
},
{
"decision_loss": 2.5931,
"encoder_mlm_loss": 0.5477,
"epoch": 0.44,
"grad_norm": 10.537424087524414,
"learning_rate": 3.770839121978327e-05,
"loss": 4.81,
"reason_loss": NaN,
"step": 8550
},
{
"decision_loss": 2.5087,
"encoder_mlm_loss": 0.5229,
"epoch": 0.44,
"grad_norm": 10.352331161499023,
"learning_rate": 3.753473186996388e-05,
"loss": 4.7803,
"reason_loss": NaN,
"step": 8600
},
{
"decision_loss": 2.5849,
"encoder_mlm_loss": 0.5264,
"epoch": 0.45,
"grad_norm": 11.812186241149902,
"learning_rate": 3.736107252014449e-05,
"loss": 4.8995,
"reason_loss": NaN,
"step": 8650
},
{
"decision_loss": 2.5873,
"encoder_mlm_loss": 0.5366,
"epoch": 0.45,
"grad_norm": 10.932312965393066,
"learning_rate": 3.718741317032509e-05,
"loss": 4.9043,
"reason_loss": NaN,
"step": 8700
},
{
"decision_loss": 2.5046,
"encoder_mlm_loss": 0.5356,
"epoch": 0.45,
"grad_norm": 12.284061431884766,
"learning_rate": 3.70137538205057e-05,
"loss": 4.79,
"reason_loss": NaN,
"step": 8750
},
{
"decision_loss": 2.4967,
"encoder_mlm_loss": 0.5345,
"epoch": 0.45,
"grad_norm": 11.007719993591309,
"learning_rate": 3.6840094470686305e-05,
"loss": 4.7234,
"reason_loss": NaN,
"step": 8800
},
{
"decision_loss": 2.4628,
"encoder_mlm_loss": 0.5338,
"epoch": 0.46,
"grad_norm": 10.874073028564453,
"learning_rate": 3.6666435120866907e-05,
"loss": 4.7151,
"reason_loss": NaN,
"step": 8850
},
{
"decision_loss": 2.4878,
"encoder_mlm_loss": 0.5206,
"epoch": 0.46,
"grad_norm": 10.251517295837402,
"learning_rate": 3.6492775771047514e-05,
"loss": 4.7856,
"reason_loss": NaN,
"step": 8900
},
{
"decision_loss": 2.3386,
"encoder_mlm_loss": 0.5355,
"epoch": 0.46,
"grad_norm": 12.211389541625977,
"learning_rate": 3.631911642122812e-05,
"loss": 4.6161,
"reason_loss": NaN,
"step": 8950
},
{
"decision_loss": 2.385,
"encoder_mlm_loss": 0.5152,
"epoch": 0.46,
"grad_norm": 10.806251525878906,
"learning_rate": 3.614545707140873e-05,
"loss": 4.7703,
"reason_loss": NaN,
"step": 9000
},
{
"decision_loss": 2.4064,
"encoder_mlm_loss": 0.5589,
"epoch": 0.47,
"grad_norm": 10.69713020324707,
"learning_rate": 3.597179772158933e-05,
"loss": 4.6229,
"reason_loss": NaN,
"step": 9050
},
{
"decision_loss": 2.5413,
"encoder_mlm_loss": 0.5165,
"epoch": 0.47,
"grad_norm": 10.956335067749023,
"learning_rate": 3.579813837176994e-05,
"loss": 4.8993,
"reason_loss": NaN,
"step": 9100
},
{
"decision_loss": 2.4698,
"encoder_mlm_loss": 0.5077,
"epoch": 0.47,
"grad_norm": 10.868124961853027,
"learning_rate": 3.562447902195054e-05,
"loss": 4.6673,
"reason_loss": NaN,
"step": 9150
},
{
"decision_loss": 2.4918,
"encoder_mlm_loss": 0.5106,
"epoch": 0.47,
"grad_norm": 10.498440742492676,
"learning_rate": 3.545081967213115e-05,
"loss": 4.7566,
"reason_loss": NaN,
"step": 9200
},
{
"decision_loss": 2.4684,
"encoder_mlm_loss": 0.5478,
"epoch": 0.48,
"grad_norm": 9.486736297607422,
"learning_rate": 3.5277160322311756e-05,
"loss": 4.7814,
"reason_loss": NaN,
"step": 9250
},
{
"decision_loss": 2.4031,
"encoder_mlm_loss": 0.5311,
"epoch": 0.48,
"grad_norm": 10.80704116821289,
"learning_rate": 3.510350097249236e-05,
"loss": 4.6901,
"reason_loss": NaN,
"step": 9300
},
{
"decision_loss": 2.4332,
"encoder_mlm_loss": 0.5448,
"epoch": 0.48,
"grad_norm": 13.817115783691406,
"learning_rate": 3.4929841622672965e-05,
"loss": 4.5975,
"reason_loss": NaN,
"step": 9350
},
{
"decision_loss": 2.5099,
"encoder_mlm_loss": 0.5268,
"epoch": 0.48,
"grad_norm": 11.08284854888916,
"learning_rate": 3.475618227285357e-05,
"loss": 4.7289,
"reason_loss": NaN,
"step": 9400
},
{
"decision_loss": 2.3983,
"encoder_mlm_loss": 0.5269,
"epoch": 0.49,
"grad_norm": 10.758065223693848,
"learning_rate": 3.4582522923034174e-05,
"loss": 4.5931,
"reason_loss": NaN,
"step": 9450
},
{
"decision_loss": 2.4453,
"encoder_mlm_loss": 0.539,
"epoch": 0.49,
"grad_norm": 12.486990928649902,
"learning_rate": 3.440886357321478e-05,
"loss": 4.7225,
"reason_loss": NaN,
"step": 9500
},
{
"decision_loss": 2.4529,
"encoder_mlm_loss": 0.5225,
"epoch": 0.49,
"grad_norm": 10.317234992980957,
"learning_rate": 3.423520422339539e-05,
"loss": 4.6519,
"reason_loss": NaN,
"step": 9550
},
{
"decision_loss": 2.5537,
"encoder_mlm_loss": 0.5198,
"epoch": 0.49,
"grad_norm": 10.231575012207031,
"learning_rate": 3.4061544873576e-05,
"loss": 4.6996,
"reason_loss": NaN,
"step": 9600
},
{
"decision_loss": 2.4312,
"encoder_mlm_loss": 0.5115,
"epoch": 0.5,
"grad_norm": 11.516695976257324,
"learning_rate": 3.38878855237566e-05,
"loss": 4.7901,
"reason_loss": NaN,
"step": 9650
},
{
"decision_loss": 2.3535,
"encoder_mlm_loss": 0.5464,
"epoch": 0.5,
"grad_norm": 10.423235893249512,
"learning_rate": 3.371422617393721e-05,
"loss": 4.67,
"reason_loss": NaN,
"step": 9700
},
{
"decision_loss": 2.3405,
"encoder_mlm_loss": 0.5209,
"epoch": 0.5,
"grad_norm": 10.108226776123047,
"learning_rate": 3.3540566824117815e-05,
"loss": 4.6348,
"reason_loss": NaN,
"step": 9750
},
{
"decision_loss": 2.4064,
"encoder_mlm_loss": 0.5045,
"epoch": 0.51,
"grad_norm": 12.5283203125,
"learning_rate": 3.336690747429842e-05,
"loss": 4.6198,
"reason_loss": NaN,
"step": 9800
},
{
"decision_loss": 2.3878,
"encoder_mlm_loss": 0.4968,
"epoch": 0.51,
"grad_norm": 10.635274887084961,
"learning_rate": 3.3193248124479024e-05,
"loss": 4.6017,
"reason_loss": NaN,
"step": 9850
},
{
"decision_loss": 2.4227,
"encoder_mlm_loss": 0.5368,
"epoch": 0.51,
"grad_norm": 11.512763977050781,
"learning_rate": 3.3019588774659625e-05,
"loss": 4.5901,
"reason_loss": NaN,
"step": 9900
},
{
"decision_loss": 2.5284,
"encoder_mlm_loss": 0.5259,
"epoch": 0.51,
"grad_norm": 9.88230037689209,
"learning_rate": 3.284592942484023e-05,
"loss": 4.5671,
"reason_loss": NaN,
"step": 9950
},
{
"decision_loss": 2.3601,
"encoder_mlm_loss": 0.5258,
"epoch": 0.52,
"grad_norm": 11.324311256408691,
"learning_rate": 3.267227007502084e-05,
"loss": 4.5307,
"reason_loss": NaN,
"step": 10000
},
{
"decision_loss": 2.2871,
"encoder_mlm_loss": 0.5113,
"epoch": 0.52,
"grad_norm": 11.137092590332031,
"learning_rate": 3.249861072520144e-05,
"loss": 4.6016,
"reason_loss": NaN,
"step": 10050
},
{
"decision_loss": 2.4347,
"encoder_mlm_loss": 0.5186,
"epoch": 0.52,
"grad_norm": 12.242900848388672,
"learning_rate": 3.232495137538205e-05,
"loss": 4.697,
"reason_loss": NaN,
"step": 10100
},
{
"decision_loss": 2.3998,
"encoder_mlm_loss": 0.5311,
"epoch": 0.52,
"grad_norm": 12.156739234924316,
"learning_rate": 3.215129202556266e-05,
"loss": 4.5704,
"reason_loss": NaN,
"step": 10150
},
{
"decision_loss": 2.5238,
"encoder_mlm_loss": 0.5182,
"epoch": 0.53,
"grad_norm": 9.229127883911133,
"learning_rate": 3.1977632675743266e-05,
"loss": 4.5861,
"reason_loss": NaN,
"step": 10200
},
{
"decision_loss": 2.378,
"encoder_mlm_loss": 0.5352,
"epoch": 0.53,
"grad_norm": 10.308478355407715,
"learning_rate": 3.180397332592387e-05,
"loss": 4.5989,
"reason_loss": NaN,
"step": 10250
},
{
"decision_loss": 2.4153,
"encoder_mlm_loss": 0.4822,
"epoch": 0.53,
"grad_norm": 10.68014144897461,
"learning_rate": 3.1630313976104475e-05,
"loss": 4.5126,
"reason_loss": NaN,
"step": 10300
},
{
"decision_loss": 2.4028,
"encoder_mlm_loss": 0.5217,
"epoch": 0.53,
"grad_norm": 10.47494125366211,
"learning_rate": 3.145665462628508e-05,
"loss": 4.5665,
"reason_loss": NaN,
"step": 10350
},
{
"decision_loss": 2.3192,
"encoder_mlm_loss": 0.5054,
"epoch": 0.54,
"grad_norm": 12.187615394592285,
"learning_rate": 3.128299527646569e-05,
"loss": 4.4394,
"reason_loss": NaN,
"step": 10400
},
{
"decision_loss": 2.3345,
"encoder_mlm_loss": 0.514,
"epoch": 0.54,
"grad_norm": 10.271819114685059,
"learning_rate": 3.110933592664629e-05,
"loss": 4.574,
"reason_loss": NaN,
"step": 10450
},
{
"decision_loss": 2.1869,
"encoder_mlm_loss": 0.49,
"epoch": 0.54,
"grad_norm": 10.244134902954102,
"learning_rate": 3.09356765768269e-05,
"loss": 4.3925,
"reason_loss": NaN,
"step": 10500
},
{
"decision_loss": 2.5201,
"encoder_mlm_loss": 0.4986,
"epoch": 0.54,
"grad_norm": 9.698922157287598,
"learning_rate": 3.076201722700751e-05,
"loss": 4.5421,
"reason_loss": NaN,
"step": 10550
},
{
"decision_loss": 2.3565,
"encoder_mlm_loss": 0.4935,
"epoch": 0.55,
"grad_norm": 10.40183162689209,
"learning_rate": 3.058835787718811e-05,
"loss": 4.5004,
"reason_loss": NaN,
"step": 10600
},
{
"decision_loss": 2.2853,
"encoder_mlm_loss": 0.4831,
"epoch": 0.55,
"grad_norm": 10.580133438110352,
"learning_rate": 3.0414698527368713e-05,
"loss": 4.5389,
"reason_loss": NaN,
"step": 10650
},
{
"decision_loss": 2.3804,
"encoder_mlm_loss": 0.4932,
"epoch": 0.55,
"grad_norm": 10.064751625061035,
"learning_rate": 3.0241039177549318e-05,
"loss": 4.5354,
"reason_loss": NaN,
"step": 10700
},
{
"decision_loss": 2.4321,
"encoder_mlm_loss": 0.4953,
"epoch": 0.55,
"grad_norm": 11.207749366760254,
"learning_rate": 3.0067379827729926e-05,
"loss": 4.5193,
"reason_loss": NaN,
"step": 10750
},
{
"decision_loss": 2.4432,
"encoder_mlm_loss": 0.5103,
"epoch": 0.56,
"grad_norm": 9.562568664550781,
"learning_rate": 2.989372047791053e-05,
"loss": 4.4666,
"reason_loss": NaN,
"step": 10800
},
{
"decision_loss": 2.3398,
"encoder_mlm_loss": 0.4874,
"epoch": 0.56,
"grad_norm": 10.37126350402832,
"learning_rate": 2.9720061128091138e-05,
"loss": 4.401,
"reason_loss": 1.3739,
"step": 10850
},
{
"decision_loss": 2.2467,
"encoder_mlm_loss": 0.4961,
"epoch": 0.56,
"grad_norm": 10.654061317443848,
"learning_rate": 2.9546401778271743e-05,
"loss": 4.5209,
"reason_loss": NaN,
"step": 10900
},
{
"decision_loss": 2.3339,
"encoder_mlm_loss": 0.5158,
"epoch": 0.56,
"grad_norm": 9.291821479797363,
"learning_rate": 2.937274242845235e-05,
"loss": 4.5287,
"reason_loss": NaN,
"step": 10950
},
{
"decision_loss": 2.4592,
"encoder_mlm_loss": 0.504,
"epoch": 0.57,
"grad_norm": 21.14087677001953,
"learning_rate": 2.9199083078632955e-05,
"loss": 4.4862,
"reason_loss": NaN,
"step": 11000
},
{
"decision_loss": 2.4528,
"encoder_mlm_loss": 0.5224,
"epoch": 0.57,
"grad_norm": 14.19636058807373,
"learning_rate": 2.902542372881356e-05,
"loss": 4.5173,
"reason_loss": NaN,
"step": 11050
},
{
"decision_loss": 2.227,
"encoder_mlm_loss": 0.5062,
"epoch": 0.57,
"grad_norm": 10.437397956848145,
"learning_rate": 2.8851764378994167e-05,
"loss": 4.416,
"reason_loss": NaN,
"step": 11100
},
{
"decision_loss": 2.3802,
"encoder_mlm_loss": 0.4667,
"epoch": 0.57,
"grad_norm": 10.666945457458496,
"learning_rate": 2.8678105029174772e-05,
"loss": 4.5593,
"reason_loss": NaN,
"step": 11150
},
{
"decision_loss": 2.4548,
"encoder_mlm_loss": 0.5221,
"epoch": 0.58,
"grad_norm": 10.959080696105957,
"learning_rate": 2.850444567935538e-05,
"loss": 4.363,
"reason_loss": NaN,
"step": 11200
},
{
"decision_loss": 2.3382,
"encoder_mlm_loss": 0.4761,
"epoch": 0.58,
"grad_norm": 9.787275314331055,
"learning_rate": 2.8330786329535984e-05,
"loss": 4.5643,
"reason_loss": NaN,
"step": 11250
},
{
"decision_loss": 2.3277,
"encoder_mlm_loss": 0.4821,
"epoch": 0.58,
"grad_norm": 9.431136131286621,
"learning_rate": 2.8157126979716592e-05,
"loss": 4.4429,
"reason_loss": NaN,
"step": 11300
},
{
"decision_loss": 2.3421,
"encoder_mlm_loss": 0.4861,
"epoch": 0.59,
"grad_norm": 9.479231834411621,
"learning_rate": 2.7983467629897197e-05,
"loss": 4.4545,
"reason_loss": NaN,
"step": 11350
},
{
"decision_loss": 2.2753,
"encoder_mlm_loss": 0.5042,
"epoch": 0.59,
"grad_norm": 11.047704696655273,
"learning_rate": 2.7813281467074186e-05,
"loss": 4.3748,
"reason_loss": NaN,
"step": 11400
},
{
"decision_loss": 2.2527,
"encoder_mlm_loss": 0.5045,
"epoch": 0.59,
"grad_norm": 11.358301162719727,
"learning_rate": 2.7639622117254794e-05,
"loss": 4.5086,
"reason_loss": NaN,
"step": 11450
},
{
"decision_loss": 2.3986,
"encoder_mlm_loss": 0.5054,
"epoch": 0.59,
"grad_norm": 12.649896621704102,
"learning_rate": 2.7465962767435398e-05,
"loss": 4.394,
"reason_loss": NaN,
"step": 11500
},
{
"decision_loss": 2.3123,
"encoder_mlm_loss": 0.5175,
"epoch": 0.6,
"grad_norm": 17.416446685791016,
"learning_rate": 2.7292303417616006e-05,
"loss": 4.4608,
"reason_loss": NaN,
"step": 11550
},
{
"decision_loss": 2.3739,
"encoder_mlm_loss": 0.4938,
"epoch": 0.6,
"grad_norm": 10.390993118286133,
"learning_rate": 2.711864406779661e-05,
"loss": 4.3458,
"reason_loss": NaN,
"step": 11600
},
{
"decision_loss": 2.3226,
"encoder_mlm_loss": 0.4896,
"epoch": 0.6,
"grad_norm": 11.228699684143066,
"learning_rate": 2.694498471797722e-05,
"loss": 4.505,
"reason_loss": NaN,
"step": 11650
},
{
"decision_loss": 2.2692,
"encoder_mlm_loss": 0.502,
"epoch": 0.6,
"grad_norm": 10.9651517868042,
"learning_rate": 2.6771325368157823e-05,
"loss": 4.3374,
"reason_loss": NaN,
"step": 11700
},
{
"decision_loss": 2.252,
"encoder_mlm_loss": 0.4922,
"epoch": 0.61,
"grad_norm": 16.753633499145508,
"learning_rate": 2.659766601833843e-05,
"loss": 4.332,
"reason_loss": NaN,
"step": 11750
},
{
"decision_loss": 2.3924,
"encoder_mlm_loss": 0.5105,
"epoch": 0.61,
"grad_norm": 11.315303802490234,
"learning_rate": 2.6424006668519035e-05,
"loss": 4.3855,
"reason_loss": NaN,
"step": 11800
},
{
"decision_loss": 2.2546,
"encoder_mlm_loss": 0.5138,
"epoch": 0.61,
"grad_norm": 10.216447830200195,
"learning_rate": 2.6250347318699643e-05,
"loss": 4.4484,
"reason_loss": NaN,
"step": 11850
},
{
"decision_loss": 2.4346,
"encoder_mlm_loss": 0.524,
"epoch": 0.61,
"grad_norm": 10.381539344787598,
"learning_rate": 2.6076687968880248e-05,
"loss": 4.4378,
"reason_loss": NaN,
"step": 11900
},
{
"decision_loss": 2.4519,
"encoder_mlm_loss": 0.5054,
"epoch": 0.62,
"grad_norm": 15.222419738769531,
"learning_rate": 2.5903028619060856e-05,
"loss": 4.5089,
"reason_loss": NaN,
"step": 11950
},
{
"decision_loss": 2.3408,
"encoder_mlm_loss": 0.5097,
"epoch": 0.62,
"grad_norm": 10.49439525604248,
"learning_rate": 2.5729369269241453e-05,
"loss": 4.3041,
"reason_loss": NaN,
"step": 12000
},
{
"decision_loss": 2.3097,
"encoder_mlm_loss": 0.486,
"epoch": 0.62,
"grad_norm": 10.5269136428833,
"learning_rate": 2.555570991942206e-05,
"loss": 4.3974,
"reason_loss": NaN,
"step": 12050
},
{
"decision_loss": 2.3363,
"encoder_mlm_loss": 0.517,
"epoch": 0.62,
"grad_norm": 10.9442138671875,
"learning_rate": 2.5382050569602666e-05,
"loss": 4.3362,
"reason_loss": NaN,
"step": 12100
},
{
"decision_loss": 2.2017,
"encoder_mlm_loss": 0.5125,
"epoch": 0.63,
"grad_norm": 10.069433212280273,
"learning_rate": 2.5208391219783274e-05,
"loss": 4.4084,
"reason_loss": NaN,
"step": 12150
},
{
"decision_loss": 2.3537,
"encoder_mlm_loss": 0.5052,
"epoch": 0.63,
"grad_norm": 10.322944641113281,
"learning_rate": 2.5034731869963878e-05,
"loss": 4.4116,
"reason_loss": NaN,
"step": 12200
},
{
"decision_loss": 2.4167,
"encoder_mlm_loss": 0.5104,
"epoch": 0.63,
"grad_norm": 9.33908462524414,
"learning_rate": 2.4861072520144486e-05,
"loss": 4.4243,
"reason_loss": NaN,
"step": 12250
},
{
"decision_loss": 2.3596,
"encoder_mlm_loss": 0.4972,
"epoch": 0.63,
"grad_norm": 21.888282775878906,
"learning_rate": 2.468741317032509e-05,
"loss": 4.3607,
"reason_loss": NaN,
"step": 12300
},
{
"decision_loss": 2.3476,
"encoder_mlm_loss": 0.4891,
"epoch": 0.64,
"grad_norm": 10.44153118133545,
"learning_rate": 2.45137538205057e-05,
"loss": 4.4001,
"reason_loss": NaN,
"step": 12350
},
{
"decision_loss": 2.2677,
"encoder_mlm_loss": 0.4686,
"epoch": 0.64,
"grad_norm": 10.878878593444824,
"learning_rate": 2.4340094470686303e-05,
"loss": 4.2944,
"reason_loss": NaN,
"step": 12400
},
{
"decision_loss": 2.3666,
"encoder_mlm_loss": 0.495,
"epoch": 0.64,
"grad_norm": 10.5595121383667,
"learning_rate": 2.4166435120866908e-05,
"loss": 4.243,
"reason_loss": NaN,
"step": 12450
},
{
"decision_loss": 2.2113,
"encoder_mlm_loss": 0.4877,
"epoch": 0.64,
"grad_norm": 10.240670204162598,
"learning_rate": 2.3992775771047516e-05,
"loss": 4.321,
"reason_loss": NaN,
"step": 12500
},
{
"decision_loss": 2.2091,
"encoder_mlm_loss": 0.5136,
"epoch": 0.65,
"grad_norm": 10.264183044433594,
"learning_rate": 2.381911642122812e-05,
"loss": 4.5137,
"reason_loss": NaN,
"step": 12550
},
{
"decision_loss": 2.3554,
"encoder_mlm_loss": 0.4939,
"epoch": 0.65,
"grad_norm": 9.599092483520508,
"learning_rate": 2.3645457071408725e-05,
"loss": 4.1951,
"reason_loss": NaN,
"step": 12600
},
{
"decision_loss": 2.3426,
"encoder_mlm_loss": 0.5021,
"epoch": 0.65,
"grad_norm": 10.048467636108398,
"learning_rate": 2.3471797721589332e-05,
"loss": 4.3319,
"reason_loss": NaN,
"step": 12650
},
{
"decision_loss": 2.4104,
"encoder_mlm_loss": 0.5072,
"epoch": 0.65,
"grad_norm": 9.640985488891602,
"learning_rate": 2.3298138371769937e-05,
"loss": 4.4168,
"reason_loss": NaN,
"step": 12700
},
{
"decision_loss": 2.3776,
"encoder_mlm_loss": 0.4877,
"epoch": 0.66,
"grad_norm": 11.355530738830566,
"learning_rate": 2.312447902195054e-05,
"loss": 4.4094,
"reason_loss": NaN,
"step": 12750
},
{
"decision_loss": 2.2999,
"encoder_mlm_loss": 0.506,
"epoch": 0.66,
"grad_norm": 9.723401069641113,
"learning_rate": 2.295081967213115e-05,
"loss": 4.3139,
"reason_loss": NaN,
"step": 12800
},
{
"decision_loss": 2.1432,
"encoder_mlm_loss": 0.4898,
"epoch": 0.66,
"grad_norm": 9.709599494934082,
"learning_rate": 2.2777160322311754e-05,
"loss": 4.3763,
"reason_loss": NaN,
"step": 12850
},
{
"decision_loss": 2.2003,
"encoder_mlm_loss": 0.4835,
"epoch": 0.67,
"grad_norm": 10.115660667419434,
"learning_rate": 2.2603500972492362e-05,
"loss": 4.262,
"reason_loss": NaN,
"step": 12900
},
{
"decision_loss": 2.2885,
"encoder_mlm_loss": 0.5165,
"epoch": 0.67,
"grad_norm": 9.499692916870117,
"learning_rate": 2.2429841622672963e-05,
"loss": 4.3976,
"reason_loss": NaN,
"step": 12950
},
{
"decision_loss": 2.3993,
"encoder_mlm_loss": 0.49,
"epoch": 0.67,
"grad_norm": 10.184179306030273,
"learning_rate": 2.225618227285357e-05,
"loss": 4.3492,
"reason_loss": NaN,
"step": 13000
},
{
"decision_loss": 2.3194,
"encoder_mlm_loss": 0.5021,
"epoch": 0.67,
"grad_norm": 10.164281845092773,
"learning_rate": 2.2082522923034175e-05,
"loss": 4.2254,
"reason_loss": NaN,
"step": 13050
},
{
"decision_loss": 2.202,
"encoder_mlm_loss": 0.4979,
"epoch": 0.68,
"grad_norm": 10.495577812194824,
"learning_rate": 2.1908863573214783e-05,
"loss": 4.2881,
"reason_loss": NaN,
"step": 13100
},
{
"decision_loss": 2.3032,
"encoder_mlm_loss": 0.5051,
"epoch": 0.68,
"grad_norm": 11.644469261169434,
"learning_rate": 2.1735204223395388e-05,
"loss": 4.2034,
"reason_loss": NaN,
"step": 13150
},
{
"decision_loss": 2.2369,
"encoder_mlm_loss": 0.494,
"epoch": 0.68,
"grad_norm": 10.125290870666504,
"learning_rate": 2.1561544873575996e-05,
"loss": 4.2488,
"reason_loss": NaN,
"step": 13200
},
{
"decision_loss": 2.2961,
"encoder_mlm_loss": 0.4633,
"epoch": 0.68,
"grad_norm": 9.7282075881958,
"learning_rate": 2.13878855237566e-05,
"loss": 4.3103,
"reason_loss": NaN,
"step": 13250
},
{
"decision_loss": 2.2269,
"encoder_mlm_loss": 0.5088,
"epoch": 0.69,
"grad_norm": 10.021200180053711,
"learning_rate": 2.1214226173937205e-05,
"loss": 4.2561,
"reason_loss": NaN,
"step": 13300
},
{
"decision_loss": 2.2716,
"encoder_mlm_loss": 0.4972,
"epoch": 0.69,
"grad_norm": 9.810166358947754,
"learning_rate": 2.104056682411781e-05,
"loss": 4.2693,
"reason_loss": NaN,
"step": 13350
},
{
"decision_loss": 2.4064,
"encoder_mlm_loss": 0.5313,
"epoch": 0.69,
"grad_norm": 9.662689208984375,
"learning_rate": 2.0866907474298417e-05,
"loss": 4.2489,
"reason_loss": NaN,
"step": 13400
},
{
"decision_loss": 2.2747,
"encoder_mlm_loss": 0.4795,
"epoch": 0.69,
"grad_norm": 9.8899507522583,
"learning_rate": 2.0693248124479022e-05,
"loss": 4.3083,
"reason_loss": NaN,
"step": 13450
},
{
"decision_loss": 2.1515,
"encoder_mlm_loss": 0.4694,
"epoch": 0.7,
"grad_norm": 10.865926742553711,
"learning_rate": 2.051958877465963e-05,
"loss": 4.2286,
"reason_loss": NaN,
"step": 13500
},
{
"decision_loss": 2.2069,
"encoder_mlm_loss": 0.4688,
"epoch": 0.7,
"grad_norm": 11.5147123336792,
"learning_rate": 2.0345929424840234e-05,
"loss": 4.2459,
"reason_loss": NaN,
"step": 13550
},
{
"decision_loss": 2.196,
"encoder_mlm_loss": 0.4853,
"epoch": 0.7,
"grad_norm": 10.14575481414795,
"learning_rate": 2.0172270075020842e-05,
"loss": 4.2824,
"reason_loss": NaN,
"step": 13600
},
{
"decision_loss": 2.0801,
"encoder_mlm_loss": 0.4571,
"epoch": 0.7,
"grad_norm": 10.563467979431152,
"learning_rate": 1.9998610725201447e-05,
"loss": 4.2221,
"reason_loss": NaN,
"step": 13650
},
{
"decision_loss": 2.2988,
"encoder_mlm_loss": 0.5028,
"epoch": 0.71,
"grad_norm": 11.101655006408691,
"learning_rate": 1.982495137538205e-05,
"loss": 4.3191,
"reason_loss": NaN,
"step": 13700
},
{
"decision_loss": 2.1795,
"encoder_mlm_loss": 0.4518,
"epoch": 0.71,
"grad_norm": 9.590519905090332,
"learning_rate": 1.9651292025562656e-05,
"loss": 4.2817,
"reason_loss": NaN,
"step": 13750
},
{
"decision_loss": 2.2984,
"encoder_mlm_loss": 0.4861,
"epoch": 0.71,
"grad_norm": 9.86621379852295,
"learning_rate": 1.9477632675743263e-05,
"loss": 4.3165,
"reason_loss": NaN,
"step": 13800
},
{
"decision_loss": 2.36,
"encoder_mlm_loss": 0.4777,
"epoch": 0.71,
"grad_norm": 10.772181510925293,
"learning_rate": 1.9303973325923868e-05,
"loss": 4.2868,
"reason_loss": NaN,
"step": 13850
},
{
"decision_loss": 2.3243,
"encoder_mlm_loss": 0.4844,
"epoch": 0.72,
"grad_norm": 11.583396911621094,
"learning_rate": 1.9130313976104476e-05,
"loss": 4.2421,
"reason_loss": NaN,
"step": 13900
},
{
"decision_loss": 2.1823,
"encoder_mlm_loss": 0.4749,
"epoch": 0.72,
"grad_norm": 10.409939765930176,
"learning_rate": 1.895665462628508e-05,
"loss": 4.2622,
"reason_loss": NaN,
"step": 13950
},
{
"decision_loss": 2.1868,
"encoder_mlm_loss": 0.5075,
"epoch": 0.72,
"grad_norm": 9.551398277282715,
"learning_rate": 1.878299527646569e-05,
"loss": 4.2983,
"reason_loss": NaN,
"step": 14000
},
{
"decision_loss": 2.1898,
"encoder_mlm_loss": 0.4868,
"epoch": 0.72,
"grad_norm": 10.858856201171875,
"learning_rate": 1.860933592664629e-05,
"loss": 4.2047,
"reason_loss": NaN,
"step": 14050
},
{
"decision_loss": 2.3408,
"encoder_mlm_loss": 0.4726,
"epoch": 0.73,
"grad_norm": 9.92735481262207,
"learning_rate": 1.8435676576826897e-05,
"loss": 4.3045,
"reason_loss": NaN,
"step": 14100
},
{
"decision_loss": 2.1705,
"encoder_mlm_loss": 0.4829,
"epoch": 0.73,
"grad_norm": 9.894718170166016,
"learning_rate": 1.8262017227007502e-05,
"loss": 4.1323,
"reason_loss": NaN,
"step": 14150
},
{
"decision_loss": 2.2165,
"encoder_mlm_loss": 0.4986,
"epoch": 0.73,
"grad_norm": 9.643003463745117,
"learning_rate": 1.808835787718811e-05,
"loss": 4.1831,
"reason_loss": NaN,
"step": 14200
},
{
"decision_loss": 2.219,
"encoder_mlm_loss": 0.5092,
"epoch": 0.73,
"grad_norm": 10.380711555480957,
"learning_rate": 1.7914698527368714e-05,
"loss": 4.2177,
"reason_loss": NaN,
"step": 14250
},
{
"decision_loss": 2.2572,
"encoder_mlm_loss": 0.4603,
"epoch": 0.74,
"grad_norm": 10.050085067749023,
"learning_rate": 1.7741039177549322e-05,
"loss": 4.2955,
"reason_loss": NaN,
"step": 14300
},
{
"decision_loss": 2.3477,
"encoder_mlm_loss": 0.4771,
"epoch": 0.74,
"grad_norm": 11.214825630187988,
"learning_rate": 1.7567379827729927e-05,
"loss": 4.1521,
"reason_loss": NaN,
"step": 14350
},
{
"decision_loss": 2.1234,
"encoder_mlm_loss": 0.4793,
"epoch": 0.74,
"grad_norm": 10.145916938781738,
"learning_rate": 1.739372047791053e-05,
"loss": 4.1726,
"reason_loss": NaN,
"step": 14400
},
{
"decision_loss": 2.2479,
"encoder_mlm_loss": 0.4941,
"epoch": 0.74,
"grad_norm": 11.086326599121094,
"learning_rate": 1.7220061128091136e-05,
"loss": 4.2883,
"reason_loss": NaN,
"step": 14450
},
{
"decision_loss": 2.0809,
"encoder_mlm_loss": 0.4895,
"epoch": 0.75,
"grad_norm": 9.647579193115234,
"learning_rate": 1.7046401778271744e-05,
"loss": 4.0984,
"reason_loss": NaN,
"step": 14500
},
{
"decision_loss": 2.2474,
"encoder_mlm_loss": 0.4863,
"epoch": 0.75,
"grad_norm": 9.199592590332031,
"learning_rate": 1.6872742428452348e-05,
"loss": 4.145,
"reason_loss": NaN,
"step": 14550
},
{
"decision_loss": 2.2259,
"encoder_mlm_loss": 0.467,
"epoch": 0.75,
"grad_norm": 10.98985767364502,
"learning_rate": 1.6699083078632953e-05,
"loss": 4.2657,
"reason_loss": NaN,
"step": 14600
},
{
"decision_loss": 2.3778,
"encoder_mlm_loss": 0.4754,
"epoch": 0.76,
"grad_norm": 12.066682815551758,
"learning_rate": 1.652542372881356e-05,
"loss": 4.2173,
"reason_loss": NaN,
"step": 14650
},
{
"decision_loss": 2.292,
"encoder_mlm_loss": 0.4871,
"epoch": 0.76,
"grad_norm": 11.109633445739746,
"learning_rate": 1.6351764378994165e-05,
"loss": 4.2824,
"reason_loss": NaN,
"step": 14700
},
{
"decision_loss": 2.3616,
"encoder_mlm_loss": 0.4859,
"epoch": 0.76,
"grad_norm": 10.090713500976562,
"learning_rate": 1.6178105029174773e-05,
"loss": 4.2862,
"reason_loss": NaN,
"step": 14750
},
{
"decision_loss": 2.23,
"encoder_mlm_loss": 0.488,
"epoch": 0.76,
"grad_norm": 12.84520149230957,
"learning_rate": 1.6004445679355378e-05,
"loss": 4.3239,
"reason_loss": NaN,
"step": 14800
},
{
"decision_loss": 2.199,
"encoder_mlm_loss": 0.4561,
"epoch": 0.77,
"grad_norm": 13.205440521240234,
"learning_rate": 1.5830786329535982e-05,
"loss": 4.2294,
"reason_loss": NaN,
"step": 14850
},
{
"decision_loss": 2.2446,
"encoder_mlm_loss": 0.4848,
"epoch": 0.77,
"grad_norm": 10.617562294006348,
"learning_rate": 1.5657126979716587e-05,
"loss": 4.1444,
"reason_loss": NaN,
"step": 14900
},
{
"decision_loss": 2.2661,
"encoder_mlm_loss": 0.4872,
"epoch": 0.77,
"grad_norm": 10.414337158203125,
"learning_rate": 1.5483467629897195e-05,
"loss": 4.0975,
"reason_loss": NaN,
"step": 14950
},
{
"decision_loss": 2.1838,
"encoder_mlm_loss": 0.4605,
"epoch": 0.77,
"grad_norm": 9.546090126037598,
"learning_rate": 1.53098082800778e-05,
"loss": 4.2057,
"reason_loss": NaN,
"step": 15000
},
{
"decision_loss": 2.284,
"encoder_mlm_loss": 0.4937,
"epoch": 0.78,
"grad_norm": 10.734001159667969,
"learning_rate": 1.5136148930258407e-05,
"loss": 4.204,
"reason_loss": NaN,
"step": 15050
},
{
"decision_loss": 2.2667,
"encoder_mlm_loss": 0.4813,
"epoch": 0.78,
"grad_norm": 11.810020446777344,
"learning_rate": 1.4962489580439013e-05,
"loss": 4.3455,
"reason_loss": NaN,
"step": 15100
},
{
"decision_loss": 2.272,
"encoder_mlm_loss": 0.4816,
"epoch": 0.78,
"grad_norm": 9.174554824829102,
"learning_rate": 1.478883023061962e-05,
"loss": 4.3054,
"reason_loss": NaN,
"step": 15150
},
{
"decision_loss": 2.2562,
"encoder_mlm_loss": 0.4649,
"epoch": 0.78,
"grad_norm": 10.114463806152344,
"learning_rate": 1.4615170880800222e-05,
"loss": 4.1423,
"reason_loss": NaN,
"step": 15200
},
{
"decision_loss": 2.2579,
"encoder_mlm_loss": 0.4779,
"epoch": 0.79,
"grad_norm": 9.515824317932129,
"learning_rate": 1.4441511530980828e-05,
"loss": 4.2965,
"reason_loss": NaN,
"step": 15250
},
{
"decision_loss": 2.2857,
"encoder_mlm_loss": 0.456,
"epoch": 0.79,
"grad_norm": 10.084831237792969,
"learning_rate": 1.4267852181161435e-05,
"loss": 4.1492,
"reason_loss": NaN,
"step": 15300
},
{
"decision_loss": 2.1108,
"encoder_mlm_loss": 0.4592,
"epoch": 0.79,
"grad_norm": 10.310980796813965,
"learning_rate": 1.409419283134204e-05,
"loss": 4.255,
"reason_loss": NaN,
"step": 15350
},
{
"decision_loss": 2.3223,
"encoder_mlm_loss": 0.4853,
"epoch": 0.79,
"grad_norm": 10.398266792297363,
"learning_rate": 1.3920533481522647e-05,
"loss": 4.2794,
"reason_loss": NaN,
"step": 15400
},
{
"decision_loss": 2.1705,
"encoder_mlm_loss": 0.484,
"epoch": 0.8,
"grad_norm": 9.45158576965332,
"learning_rate": 1.3746874131703252e-05,
"loss": 4.1371,
"reason_loss": NaN,
"step": 15450
},
{
"decision_loss": 2.2301,
"encoder_mlm_loss": 0.4763,
"epoch": 0.8,
"grad_norm": 10.345832824707031,
"learning_rate": 1.3573214781883858e-05,
"loss": 4.1141,
"reason_loss": NaN,
"step": 15500
},
{
"decision_loss": 2.1053,
"encoder_mlm_loss": 0.4808,
"epoch": 0.8,
"grad_norm": 10.891772270202637,
"learning_rate": 1.3399555432064462e-05,
"loss": 4.0825,
"reason_loss": NaN,
"step": 15550
},
{
"decision_loss": 2.1519,
"encoder_mlm_loss": 0.4767,
"epoch": 0.8,
"grad_norm": 9.94306755065918,
"learning_rate": 1.3225896082245068e-05,
"loss": 4.0361,
"reason_loss": NaN,
"step": 15600
},
{
"decision_loss": 2.4022,
"encoder_mlm_loss": 0.4708,
"epoch": 0.81,
"grad_norm": 10.663613319396973,
"learning_rate": 1.3052236732425675e-05,
"loss": 4.2401,
"reason_loss": NaN,
"step": 15650
},
{
"decision_loss": 2.1628,
"encoder_mlm_loss": 0.4709,
"epoch": 0.81,
"grad_norm": 10.070218086242676,
"learning_rate": 1.287857738260628e-05,
"loss": 4.098,
"reason_loss": NaN,
"step": 15700
},
{
"decision_loss": 2.1725,
"encoder_mlm_loss": 0.4752,
"epoch": 0.81,
"grad_norm": 10.272323608398438,
"learning_rate": 1.2704918032786885e-05,
"loss": 4.1846,
"reason_loss": NaN,
"step": 15750
},
{
"decision_loss": 2.1207,
"encoder_mlm_loss": 0.4822,
"epoch": 0.81,
"grad_norm": 10.824825286865234,
"learning_rate": 1.2531258682967492e-05,
"loss": 4.0792,
"reason_loss": NaN,
"step": 15800
},
{
"decision_loss": 2.1834,
"encoder_mlm_loss": 0.4447,
"epoch": 0.82,
"grad_norm": 10.454733848571777,
"learning_rate": 1.2357599333148096e-05,
"loss": 4.2325,
"reason_loss": NaN,
"step": 15850
},
{
"decision_loss": 2.164,
"encoder_mlm_loss": 0.4736,
"epoch": 0.82,
"grad_norm": 9.149787902832031,
"learning_rate": 1.2183939983328702e-05,
"loss": 4.1617,
"reason_loss": NaN,
"step": 15900
},
{
"decision_loss": 2.2075,
"encoder_mlm_loss": 0.4746,
"epoch": 0.82,
"grad_norm": 10.008898735046387,
"learning_rate": 1.2010280633509309e-05,
"loss": 4.0822,
"reason_loss": NaN,
"step": 15950
},
{
"decision_loss": 2.1779,
"encoder_mlm_loss": 0.4737,
"epoch": 0.82,
"grad_norm": 9.24558162689209,
"learning_rate": 1.1836621283689913e-05,
"loss": 4.104,
"reason_loss": NaN,
"step": 16000
},
{
"decision_loss": 2.1552,
"encoder_mlm_loss": 0.4443,
"epoch": 0.83,
"grad_norm": 10.271194458007812,
"learning_rate": 1.166296193387052e-05,
"loss": 4.1249,
"reason_loss": NaN,
"step": 16050
},
{
"decision_loss": 2.2418,
"encoder_mlm_loss": 0.476,
"epoch": 0.83,
"grad_norm": 11.877102851867676,
"learning_rate": 1.1489302584051126e-05,
"loss": 4.1547,
"reason_loss": NaN,
"step": 16100
},
{
"decision_loss": 2.0779,
"encoder_mlm_loss": 0.4786,
"epoch": 0.83,
"grad_norm": 11.429306983947754,
"learning_rate": 1.1315643234231732e-05,
"loss": 4.0876,
"reason_loss": NaN,
"step": 16150
},
{
"decision_loss": 2.178,
"encoder_mlm_loss": 0.4587,
"epoch": 0.84,
"grad_norm": 9.763303756713867,
"learning_rate": 1.1141983884412336e-05,
"loss": 4.1172,
"reason_loss": NaN,
"step": 16200
},
{
"decision_loss": 2.1244,
"encoder_mlm_loss": 0.4733,
"epoch": 0.84,
"grad_norm": 9.711063385009766,
"learning_rate": 1.0968324534592942e-05,
"loss": 4.2619,
"reason_loss": NaN,
"step": 16250
},
{
"decision_loss": 2.2836,
"encoder_mlm_loss": 0.4826,
"epoch": 0.84,
"grad_norm": 9.272294044494629,
"learning_rate": 1.0794665184773549e-05,
"loss": 4.1551,
"reason_loss": NaN,
"step": 16300
},
{
"decision_loss": 2.1646,
"encoder_mlm_loss": 0.4682,
"epoch": 0.84,
"grad_norm": 13.107208251953125,
"learning_rate": 1.0621005834954155e-05,
"loss": 4.1731,
"reason_loss": NaN,
"step": 16350
},
{
"decision_loss": 2.3081,
"encoder_mlm_loss": 0.4661,
"epoch": 0.85,
"grad_norm": 9.991193771362305,
"learning_rate": 1.044734648513476e-05,
"loss": 4.225,
"reason_loss": NaN,
"step": 16400
},
{
"decision_loss": 2.3434,
"encoder_mlm_loss": 0.4999,
"epoch": 0.85,
"grad_norm": 10.09374713897705,
"learning_rate": 1.0273687135315366e-05,
"loss": 4.1445,
"reason_loss": NaN,
"step": 16450
},
{
"decision_loss": 2.1455,
"encoder_mlm_loss": 0.4582,
"epoch": 0.85,
"grad_norm": 11.907501220703125,
"learning_rate": 1.0100027785495972e-05,
"loss": 4.0988,
"reason_loss": NaN,
"step": 16500
},
{
"decision_loss": 2.1118,
"encoder_mlm_loss": 0.456,
"epoch": 0.85,
"grad_norm": 10.339010238647461,
"learning_rate": 9.926368435676578e-06,
"loss": 4.0563,
"reason_loss": NaN,
"step": 16550
},
{
"decision_loss": 2.2369,
"encoder_mlm_loss": 0.4785,
"epoch": 0.86,
"grad_norm": 11.294181823730469,
"learning_rate": 9.752709085857183e-06,
"loss": 4.0886,
"reason_loss": NaN,
"step": 16600
},
{
"decision_loss": 2.1362,
"encoder_mlm_loss": 0.4651,
"epoch": 0.86,
"grad_norm": 9.752805709838867,
"learning_rate": 9.579049736037789e-06,
"loss": 4.2024,
"reason_loss": NaN,
"step": 16650
},
{
"decision_loss": 2.1695,
"encoder_mlm_loss": 0.4615,
"epoch": 0.86,
"grad_norm": 9.915719985961914,
"learning_rate": 9.405390386218395e-06,
"loss": 4.1516,
"reason_loss": NaN,
"step": 16700
},
{
"decision_loss": 2.1229,
"encoder_mlm_loss": 0.4496,
"epoch": 0.86,
"grad_norm": 10.295191764831543,
"learning_rate": 9.231731036399e-06,
"loss": 4.0744,
"reason_loss": NaN,
"step": 16750
},
{
"decision_loss": 2.0994,
"encoder_mlm_loss": 0.4744,
"epoch": 0.87,
"grad_norm": 9.025927543640137,
"learning_rate": 9.058071686579606e-06,
"loss": 4.1056,
"reason_loss": NaN,
"step": 16800
},
{
"decision_loss": 2.3615,
"encoder_mlm_loss": 0.4828,
"epoch": 0.87,
"grad_norm": 10.910390853881836,
"learning_rate": 8.884412336760212e-06,
"loss": 4.098,
"reason_loss": NaN,
"step": 16850
},
{
"decision_loss": 2.2499,
"encoder_mlm_loss": 0.4482,
"epoch": 0.87,
"grad_norm": 10.158785820007324,
"learning_rate": 8.710752986940818e-06,
"loss": 4.0801,
"reason_loss": NaN,
"step": 16900
},
{
"decision_loss": 2.259,
"encoder_mlm_loss": 0.4799,
"epoch": 0.87,
"grad_norm": 10.046515464782715,
"learning_rate": 8.537093637121423e-06,
"loss": 4.0939,
"reason_loss": NaN,
"step": 16950
},
{
"decision_loss": 2.311,
"encoder_mlm_loss": 0.4741,
"epoch": 0.88,
"grad_norm": 11.217142105102539,
"learning_rate": 8.363434287302029e-06,
"loss": 4.211,
"reason_loss": NaN,
"step": 17000
},
{
"decision_loss": 2.1602,
"encoder_mlm_loss": 0.483,
"epoch": 0.88,
"grad_norm": 10.031744003295898,
"learning_rate": 8.189774937482635e-06,
"loss": 4.2076,
"reason_loss": NaN,
"step": 17050
},
{
"decision_loss": 2.1981,
"encoder_mlm_loss": 0.4738,
"epoch": 0.88,
"grad_norm": 9.74100399017334,
"learning_rate": 8.016115587663241e-06,
"loss": 4.1243,
"reason_loss": NaN,
"step": 17100
},
{
"decision_loss": 2.0827,
"encoder_mlm_loss": 0.4343,
"epoch": 0.88,
"grad_norm": 10.940787315368652,
"learning_rate": 7.842456237843846e-06,
"loss": 4.1619,
"reason_loss": NaN,
"step": 17150
},
{
"decision_loss": 2.1925,
"encoder_mlm_loss": 0.4737,
"epoch": 0.89,
"grad_norm": 9.71260929107666,
"learning_rate": 7.668796888024452e-06,
"loss": 4.1807,
"reason_loss": NaN,
"step": 17200
},
{
"decision_loss": 2.013,
"encoder_mlm_loss": 0.4542,
"epoch": 0.89,
"grad_norm": 14.651382446289062,
"learning_rate": 7.495137538205057e-06,
"loss": 4.0187,
"reason_loss": NaN,
"step": 17250
},
{
"decision_loss": 2.1625,
"encoder_mlm_loss": 0.4688,
"epoch": 0.89,
"grad_norm": 9.133673667907715,
"learning_rate": 7.321478188385664e-06,
"loss": 3.9995,
"reason_loss": NaN,
"step": 17300
},
{
"decision_loss": 2.2081,
"encoder_mlm_loss": 0.4961,
"epoch": 0.89,
"grad_norm": 10.863176345825195,
"learning_rate": 7.147818838566268e-06,
"loss": 4.127,
"reason_loss": NaN,
"step": 17350
},
{
"decision_loss": 2.2125,
"encoder_mlm_loss": 0.4484,
"epoch": 0.9,
"grad_norm": 9.56596851348877,
"learning_rate": 6.974159488746874e-06,
"loss": 4.1691,
"reason_loss": NaN,
"step": 17400
},
{
"decision_loss": 1.9912,
"encoder_mlm_loss": 0.4631,
"epoch": 0.9,
"grad_norm": 9.751823425292969,
"learning_rate": 6.8005001389274806e-06,
"loss": 4.0431,
"reason_loss": NaN,
"step": 17450
},
{
"decision_loss": 2.1434,
"encoder_mlm_loss": 0.4535,
"epoch": 0.9,
"grad_norm": 11.732806205749512,
"learning_rate": 6.626840789108085e-06,
"loss": 4.1323,
"reason_loss": NaN,
"step": 17500
},
{
"decision_loss": 2.1993,
"encoder_mlm_loss": 0.447,
"epoch": 0.9,
"grad_norm": 9.667863845825195,
"learning_rate": 6.453181439288691e-06,
"loss": 4.1652,
"reason_loss": NaN,
"step": 17550
},
{
"decision_loss": 2.1847,
"encoder_mlm_loss": 0.4521,
"epoch": 0.91,
"grad_norm": 10.698508262634277,
"learning_rate": 6.2795220894692975e-06,
"loss": 3.9859,
"reason_loss": NaN,
"step": 17600
},
{
"decision_loss": 2.2392,
"encoder_mlm_loss": 0.4419,
"epoch": 0.91,
"grad_norm": 10.42066764831543,
"learning_rate": 6.105862739649903e-06,
"loss": 4.1219,
"reason_loss": 1.1305,
"step": 17650
},
{
"decision_loss": 2.1109,
"encoder_mlm_loss": 0.4651,
"epoch": 0.91,
"grad_norm": 10.604533195495605,
"learning_rate": 5.932203389830509e-06,
"loss": 4.0701,
"reason_loss": NaN,
"step": 17700
},
{
"decision_loss": 1.9828,
"encoder_mlm_loss": 0.4467,
"epoch": 0.92,
"grad_norm": 11.114700317382812,
"learning_rate": 5.7585440400111144e-06,
"loss": 4.0831,
"reason_loss": NaN,
"step": 17750
},
{
"decision_loss": 2.1991,
"encoder_mlm_loss": 0.4913,
"epoch": 0.92,
"grad_norm": 11.457265853881836,
"learning_rate": 5.584884690191721e-06,
"loss": 4.2505,
"reason_loss": NaN,
"step": 17800
},
{
"decision_loss": 2.1172,
"encoder_mlm_loss": 0.4694,
"epoch": 0.92,
"grad_norm": 10.09403133392334,
"learning_rate": 5.411225340372326e-06,
"loss": 4.0314,
"reason_loss": NaN,
"step": 17850
},
{
"decision_loss": 2.1877,
"encoder_mlm_loss": 0.4449,
"epoch": 0.92,
"grad_norm": 10.144293785095215,
"learning_rate": 5.237565990552932e-06,
"loss": 4.2075,
"reason_loss": NaN,
"step": 17900
},
{
"decision_loss": 2.2545,
"encoder_mlm_loss": 0.4551,
"epoch": 0.93,
"grad_norm": 10.994632720947266,
"learning_rate": 5.063906640733538e-06,
"loss": 4.2875,
"reason_loss": NaN,
"step": 17950
},
{
"decision_loss": 2.181,
"encoder_mlm_loss": 0.4575,
"epoch": 0.93,
"grad_norm": 13.253012657165527,
"learning_rate": 4.890247290914143e-06,
"loss": 4.1868,
"reason_loss": NaN,
"step": 18000
},
{
"decision_loss": 2.1486,
"encoder_mlm_loss": 0.457,
"epoch": 0.93,
"grad_norm": 10.766643524169922,
"learning_rate": 4.716587941094748e-06,
"loss": 4.0513,
"reason_loss": NaN,
"step": 18050
},
{
"decision_loss": 2.1202,
"encoder_mlm_loss": 0.4788,
"epoch": 0.93,
"grad_norm": 9.282386779785156,
"learning_rate": 4.5429285912753545e-06,
"loss": 4.0717,
"reason_loss": NaN,
"step": 18100
},
{
"decision_loss": 2.2307,
"encoder_mlm_loss": 0.4714,
"epoch": 0.94,
"grad_norm": 11.088687896728516,
"learning_rate": 4.36926924145596e-06,
"loss": 4.0528,
"reason_loss": NaN,
"step": 18150
},
{
"decision_loss": 2.1483,
"encoder_mlm_loss": 0.4543,
"epoch": 0.94,
"grad_norm": 10.724303245544434,
"learning_rate": 4.195609891636565e-06,
"loss": 4.1388,
"reason_loss": NaN,
"step": 18200
},
{
"decision_loss": 2.212,
"encoder_mlm_loss": 0.4605,
"epoch": 0.94,
"grad_norm": 9.722859382629395,
"learning_rate": 4.0219505418171715e-06,
"loss": 4.0295,
"reason_loss": NaN,
"step": 18250
},
{
"decision_loss": 2.2006,
"encoder_mlm_loss": 0.445,
"epoch": 0.94,
"grad_norm": 9.893216133117676,
"learning_rate": 3.848291191997777e-06,
"loss": 4.1058,
"reason_loss": NaN,
"step": 18300
},
{
"decision_loss": 2.1725,
"encoder_mlm_loss": 0.483,
"epoch": 0.95,
"grad_norm": 10.373659133911133,
"learning_rate": 3.674631842178383e-06,
"loss": 4.0683,
"reason_loss": NaN,
"step": 18350
},
{
"decision_loss": 2.3838,
"encoder_mlm_loss": 0.485,
"epoch": 0.95,
"grad_norm": 9.621594429016113,
"learning_rate": 3.5009724923589884e-06,
"loss": 4.1663,
"reason_loss": NaN,
"step": 18400
},
{
"decision_loss": 2.1828,
"encoder_mlm_loss": 0.456,
"epoch": 0.95,
"grad_norm": 10.567208290100098,
"learning_rate": 3.3273131425395946e-06,
"loss": 4.1053,
"reason_loss": NaN,
"step": 18450
},
{
"decision_loss": 2.3296,
"encoder_mlm_loss": 0.4781,
"epoch": 0.95,
"grad_norm": 9.499934196472168,
"learning_rate": 3.1536537927202e-06,
"loss": 4.0483,
"reason_loss": NaN,
"step": 18500
},
{
"decision_loss": 1.9933,
"encoder_mlm_loss": 0.4588,
"epoch": 0.96,
"grad_norm": 9.636274337768555,
"learning_rate": 2.9799944429008058e-06,
"loss": 3.9718,
"reason_loss": NaN,
"step": 18550
},
{
"decision_loss": 2.167,
"encoder_mlm_loss": 0.4472,
"epoch": 0.96,
"grad_norm": 9.905817985534668,
"learning_rate": 2.8063350930814116e-06,
"loss": 4.0502,
"reason_loss": NaN,
"step": 18600
},
{
"decision_loss": 2.2164,
"encoder_mlm_loss": 0.4483,
"epoch": 0.96,
"grad_norm": 9.581639289855957,
"learning_rate": 2.636148930258405e-06,
"loss": 4.0439,
"reason_loss": NaN,
"step": 18650
},
{
"decision_loss": 2.2307,
"encoder_mlm_loss": 0.466,
"epoch": 0.96,
"grad_norm": 10.538454055786133,
"learning_rate": 2.462489580439011e-06,
"loss": 4.079,
"reason_loss": NaN,
"step": 18700
},
{
"decision_loss": 2.1576,
"encoder_mlm_loss": 0.4452,
"epoch": 0.97,
"grad_norm": 10.088532447814941,
"learning_rate": 2.2888302306196166e-06,
"loss": 4.0738,
"reason_loss": NaN,
"step": 18750
},
{
"decision_loss": 2.1127,
"encoder_mlm_loss": 0.4568,
"epoch": 0.97,
"grad_norm": 9.137553215026855,
"learning_rate": 2.1151708808002224e-06,
"loss": 4.0466,
"reason_loss": NaN,
"step": 18800
},
{
"decision_loss": 2.1179,
"encoder_mlm_loss": 0.4432,
"epoch": 0.97,
"grad_norm": 9.486440658569336,
"learning_rate": 1.9415115309808282e-06,
"loss": 4.1585,
"reason_loss": NaN,
"step": 18850
},
{
"decision_loss": 2.1924,
"encoder_mlm_loss": 0.4469,
"epoch": 0.97,
"grad_norm": 9.941400527954102,
"learning_rate": 1.7678521811614338e-06,
"loss": 4.0744,
"reason_loss": NaN,
"step": 18900
},
{
"decision_loss": 2.1206,
"encoder_mlm_loss": 0.4569,
"epoch": 0.98,
"grad_norm": 9.793261528015137,
"learning_rate": 1.5941928313420396e-06,
"loss": 4.1067,
"reason_loss": NaN,
"step": 18950
},
{
"decision_loss": 2.0999,
"encoder_mlm_loss": 0.4246,
"epoch": 0.98,
"grad_norm": 9.890857696533203,
"learning_rate": 1.4205334815226454e-06,
"loss": 4.1567,
"reason_loss": NaN,
"step": 19000
},
{
"decision_loss": 2.1566,
"encoder_mlm_loss": 0.4574,
"epoch": 0.98,
"grad_norm": 10.175832748413086,
"learning_rate": 1.246874131703251e-06,
"loss": 4.0561,
"reason_loss": NaN,
"step": 19050
},
{
"decision_loss": 2.3035,
"encoder_mlm_loss": 0.4631,
"epoch": 0.98,
"grad_norm": 10.100351333618164,
"learning_rate": 1.0732147818838565e-06,
"loss": 4.1019,
"reason_loss": NaN,
"step": 19100
},
{
"decision_loss": 2.0727,
"encoder_mlm_loss": 0.4451,
"epoch": 0.99,
"grad_norm": 10.576125144958496,
"learning_rate": 8.995554320644623e-07,
"loss": 4.151,
"reason_loss": NaN,
"step": 19150
},
{
"decision_loss": 2.2445,
"encoder_mlm_loss": 0.4302,
"epoch": 0.99,
"grad_norm": 9.79544448852539,
"learning_rate": 7.258960822450681e-07,
"loss": 4.1227,
"reason_loss": NaN,
"step": 19200
},
{
"decision_loss": 2.2291,
"encoder_mlm_loss": 0.4474,
"epoch": 0.99,
"grad_norm": 10.398290634155273,
"learning_rate": 5.522367324256738e-07,
"loss": 3.9649,
"reason_loss": NaN,
"step": 19250
},
{
"decision_loss": 2.2452,
"encoder_mlm_loss": 0.4784,
"epoch": 1.0,
"grad_norm": 11.05727481842041,
"learning_rate": 3.7857738260627953e-07,
"loss": 4.0789,
"reason_loss": NaN,
"step": 19300
},
{
"decision_loss": 2.2477,
"encoder_mlm_loss": 0.4668,
"epoch": 1.0,
"grad_norm": 10.685857772827148,
"learning_rate": 2.0491803278688526e-07,
"loss": 4.0863,
"reason_loss": NaN,
"step": 19350
}
],
"logging_steps": 50,
"max_steps": 19396,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 5000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.0030959011014115e+19,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}