| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.13541489903275072, |
| "eval_steps": 500, |
| "global_step": 133, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.001018157135584592, |
| "grad_norm": 11.6351900100708, |
| "learning_rate": 0.0, |
| "loss": 9.5408, |
| "num_input_tokens_seen": 1572864, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.002036314271169184, |
| "grad_norm": 12.06759262084961, |
| "learning_rate": 3.7037037037037037e-06, |
| "loss": 9.2981, |
| "num_input_tokens_seen": 3145728, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.0030544714067537756, |
| "grad_norm": 5.20039701461792, |
| "learning_rate": 7.4074074074074075e-06, |
| "loss": 9.1643, |
| "num_input_tokens_seen": 4718592, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.004072628542338368, |
| "grad_norm": 5.776241779327393, |
| "learning_rate": 1.1111111111111112e-05, |
| "loss": 9.155, |
| "num_input_tokens_seen": 6291456, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.0050907856779229595, |
| "grad_norm": 5.6493048667907715, |
| "learning_rate": 1.4814814814814815e-05, |
| "loss": 9.9883, |
| "num_input_tokens_seen": 7864320, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.006108942813507551, |
| "grad_norm": 5.06437873840332, |
| "learning_rate": 1.8518518518518518e-05, |
| "loss": 9.0083, |
| "num_input_tokens_seen": 9437184, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.007127099949092143, |
| "grad_norm": 8.688281059265137, |
| "learning_rate": 2.2222222222222223e-05, |
| "loss": 8.7441, |
| "num_input_tokens_seen": 11010048, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.008145257084676736, |
| "grad_norm": 4.987055778503418, |
| "learning_rate": 2.5925925925925925e-05, |
| "loss": 9.578, |
| "num_input_tokens_seen": 12582912, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.009163414220261326, |
| "grad_norm": 6.269039154052734, |
| "learning_rate": 2.962962962962963e-05, |
| "loss": 9.2803, |
| "num_input_tokens_seen": 14155776, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.010181571355845919, |
| "grad_norm": 3.3577678203582764, |
| "learning_rate": 3.3333333333333335e-05, |
| "loss": 8.7893, |
| "num_input_tokens_seen": 15728640, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.01119972849143051, |
| "grad_norm": 4.074950218200684, |
| "learning_rate": 3.7037037037037037e-05, |
| "loss": 8.8306, |
| "num_input_tokens_seen": 17301504, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.012217885627015103, |
| "grad_norm": 2.5604281425476074, |
| "learning_rate": 4.074074074074074e-05, |
| "loss": 9.9509, |
| "num_input_tokens_seen": 18874368, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.013236042762599695, |
| "grad_norm": 5.522690773010254, |
| "learning_rate": 4.4444444444444447e-05, |
| "loss": 10.0199, |
| "num_input_tokens_seen": 20447232, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.014254199898184286, |
| "grad_norm": 3.264158010482788, |
| "learning_rate": 4.814814814814815e-05, |
| "loss": 9.2761, |
| "num_input_tokens_seen": 22020096, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.015272357033768879, |
| "grad_norm": 3.1376729011535645, |
| "learning_rate": 5.185185185185185e-05, |
| "loss": 9.7989, |
| "num_input_tokens_seen": 23592960, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.01629051416935347, |
| "grad_norm": 2.6666500568389893, |
| "learning_rate": 5.555555555555556e-05, |
| "loss": 9.2875, |
| "num_input_tokens_seen": 25165824, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.017308671304938062, |
| "grad_norm": 3.044588565826416, |
| "learning_rate": 5.925925925925926e-05, |
| "loss": 8.9129, |
| "num_input_tokens_seen": 26738688, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.018326828440522653, |
| "grad_norm": 3.131470203399658, |
| "learning_rate": 6.296296296296296e-05, |
| "loss": 9.2967, |
| "num_input_tokens_seen": 28311552, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.019344985576107247, |
| "grad_norm": 3.9975271224975586, |
| "learning_rate": 6.666666666666667e-05, |
| "loss": 9.2564, |
| "num_input_tokens_seen": 29884416, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.020363142711691838, |
| "grad_norm": 4.00247049331665, |
| "learning_rate": 7.037037037037038e-05, |
| "loss": 9.7382, |
| "num_input_tokens_seen": 31457280, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.02138129984727643, |
| "grad_norm": 2.88885235786438, |
| "learning_rate": 7.407407407407407e-05, |
| "loss": 9.8961, |
| "num_input_tokens_seen": 33030144, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.02239945698286102, |
| "grad_norm": 3.772406578063965, |
| "learning_rate": 7.777777777777778e-05, |
| "loss": 10.2719, |
| "num_input_tokens_seen": 34603008, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.023417614118445614, |
| "grad_norm": 2.989847421646118, |
| "learning_rate": 8.148148148148148e-05, |
| "loss": 8.0849, |
| "num_input_tokens_seen": 36175872, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.024435771254030205, |
| "grad_norm": 3.9388954639434814, |
| "learning_rate": 8.518518518518518e-05, |
| "loss": 9.0817, |
| "num_input_tokens_seen": 37748736, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.025453928389614796, |
| "grad_norm": 3.857853889465332, |
| "learning_rate": 8.888888888888889e-05, |
| "loss": 9.2902, |
| "num_input_tokens_seen": 39321600, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.02647208552519939, |
| "grad_norm": 4.174322605133057, |
| "learning_rate": 9.25925925925926e-05, |
| "loss": 8.8552, |
| "num_input_tokens_seen": 40894464, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.02749024266078398, |
| "grad_norm": 2.5443918704986572, |
| "learning_rate": 9.62962962962963e-05, |
| "loss": 8.4878, |
| "num_input_tokens_seen": 42467328, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.028508399796368572, |
| "grad_norm": 3.078213691711426, |
| "learning_rate": 0.0001, |
| "loss": 10.4723, |
| "num_input_tokens_seen": 44040192, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.029526556931953166, |
| "grad_norm": 4.266736030578613, |
| "learning_rate": 9.999568045802217e-05, |
| "loss": 8.8477, |
| "num_input_tokens_seen": 45613056, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.030544714067537757, |
| "grad_norm": 6.914696216583252, |
| "learning_rate": 9.998272257842641e-05, |
| "loss": 8.4572, |
| "num_input_tokens_seen": 47185920, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.03156287120312235, |
| "grad_norm": 5.477447032928467, |
| "learning_rate": 9.996112860009688e-05, |
| "loss": 10.2007, |
| "num_input_tokens_seen": 48758784, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.03258102833870694, |
| "grad_norm": 4.215938091278076, |
| "learning_rate": 9.993090225407743e-05, |
| "loss": 9.4179, |
| "num_input_tokens_seen": 50331648, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.03359918547429153, |
| "grad_norm": 2.894822597503662, |
| "learning_rate": 9.989204876292688e-05, |
| "loss": 9.0953, |
| "num_input_tokens_seen": 51904512, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.034617342609876124, |
| "grad_norm": 6.472196102142334, |
| "learning_rate": 9.984457483981669e-05, |
| "loss": 9.91, |
| "num_input_tokens_seen": 53477376, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.035635499745460715, |
| "grad_norm": 2.529986619949341, |
| "learning_rate": 9.978848868737098e-05, |
| "loss": 9.3412, |
| "num_input_tokens_seen": 55050240, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.036653656881045306, |
| "grad_norm": 6.255418300628662, |
| "learning_rate": 9.972379999624936e-05, |
| "loss": 9.6125, |
| "num_input_tokens_seen": 56623104, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.0376718140166299, |
| "grad_norm": 4.211091041564941, |
| "learning_rate": 9.96505199434725e-05, |
| "loss": 9.9629, |
| "num_input_tokens_seen": 58195968, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.038689971152214495, |
| "grad_norm": 3.872631549835205, |
| "learning_rate": 9.956866119049095e-05, |
| "loss": 8.2777, |
| "num_input_tokens_seen": 59768832, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.039708128287799085, |
| "grad_norm": 2.6533520221710205, |
| "learning_rate": 9.947823788099753e-05, |
| "loss": 9.6421, |
| "num_input_tokens_seen": 61341696, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.040726285423383676, |
| "grad_norm": 2.9233040809631348, |
| "learning_rate": 9.937926563848346e-05, |
| "loss": 9.1002, |
| "num_input_tokens_seen": 62914560, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.04174444255896827, |
| "grad_norm": 2.9860687255859375, |
| "learning_rate": 9.927176156353899e-05, |
| "loss": 9.3179, |
| "num_input_tokens_seen": 64487424, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.04276259969455286, |
| "grad_norm": 8.309036254882812, |
| "learning_rate": 9.91557442308987e-05, |
| "loss": 8.7919, |
| "num_input_tokens_seen": 66060288, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.04378075683013745, |
| "grad_norm": 3.492337226867676, |
| "learning_rate": 9.903123368623216e-05, |
| "loss": 9.2796, |
| "num_input_tokens_seen": 67633152, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.04479891396572204, |
| "grad_norm": 4.4715352058410645, |
| "learning_rate": 9.889825144268029e-05, |
| "loss": 7.8785, |
| "num_input_tokens_seen": 69206016, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.04581707110130664, |
| "grad_norm": 3.341895341873169, |
| "learning_rate": 9.875682047713846e-05, |
| "loss": 8.9495, |
| "num_input_tokens_seen": 70778880, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.04683522823689123, |
| "grad_norm": 3.275592803955078, |
| "learning_rate": 9.860696522628639e-05, |
| "loss": 10.1828, |
| "num_input_tokens_seen": 72351744, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.04785338537247582, |
| "grad_norm": 3.430386781692505, |
| "learning_rate": 9.844871158236591e-05, |
| "loss": 8.193, |
| "num_input_tokens_seen": 73924608, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.04887154250806041, |
| "grad_norm": 2.59080171585083, |
| "learning_rate": 9.828208688870735e-05, |
| "loss": 8.7955, |
| "num_input_tokens_seen": 75497472, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.049889699643645, |
| "grad_norm": 3.031977653503418, |
| "learning_rate": 9.810711993500507e-05, |
| "loss": 10.1191, |
| "num_input_tokens_seen": 77070336, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.05090785677922959, |
| "grad_norm": 10.469562530517578, |
| "learning_rate": 9.792384095234313e-05, |
| "loss": 9.2337, |
| "num_input_tokens_seen": 78643200, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.05192601391481419, |
| "grad_norm": 2.8945353031158447, |
| "learning_rate": 9.773228160797188e-05, |
| "loss": 8.936, |
| "num_input_tokens_seen": 80216064, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.05294417105039878, |
| "grad_norm": 2.5451819896698, |
| "learning_rate": 9.753247499983649e-05, |
| "loss": 9.4338, |
| "num_input_tokens_seen": 81788928, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.05396232818598337, |
| "grad_norm": 2.6489815711975098, |
| "learning_rate": 9.732445565085824e-05, |
| "loss": 8.2926, |
| "num_input_tokens_seen": 83361792, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.05498048532156796, |
| "grad_norm": 2.7393391132354736, |
| "learning_rate": 9.71082595029695e-05, |
| "loss": 9.2477, |
| "num_input_tokens_seen": 84934656, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.05599864245715255, |
| "grad_norm": 2.3519155979156494, |
| "learning_rate": 9.688392391090373e-05, |
| "loss": 9.2921, |
| "num_input_tokens_seen": 86507520, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.057016799592737144, |
| "grad_norm": 3.0327858924865723, |
| "learning_rate": 9.665148763574123e-05, |
| "loss": 8.9777, |
| "num_input_tokens_seen": 88080384, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.058034956728321735, |
| "grad_norm": 3.5062544345855713, |
| "learning_rate": 9.64109908382119e-05, |
| "loss": 9.9211, |
| "num_input_tokens_seen": 89653248, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.05905311386390633, |
| "grad_norm": 2.559021472930908, |
| "learning_rate": 9.616247507175623e-05, |
| "loss": 8.9003, |
| "num_input_tokens_seen": 91226112, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.060071270999490924, |
| "grad_norm": 2.288188934326172, |
| "learning_rate": 9.590598327534564e-05, |
| "loss": 10.0134, |
| "num_input_tokens_seen": 92798976, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.061089428135075514, |
| "grad_norm": 3.029099702835083, |
| "learning_rate": 9.564155976606339e-05, |
| "loss": 9.0596, |
| "num_input_tokens_seen": 94371840, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.062107585270660105, |
| "grad_norm": 5.305948734283447, |
| "learning_rate": 9.536925023144742e-05, |
| "loss": 8.9666, |
| "num_input_tokens_seen": 95944704, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.0631257424062447, |
| "grad_norm": 2.6959710121154785, |
| "learning_rate": 9.508910172159635e-05, |
| "loss": 9.7026, |
| "num_input_tokens_seen": 97517568, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.0641438995418293, |
| "grad_norm": 2.45168399810791, |
| "learning_rate": 9.480116264104011e-05, |
| "loss": 8.1744, |
| "num_input_tokens_seen": 99090432, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.06516205667741388, |
| "grad_norm": 2.5155069828033447, |
| "learning_rate": 9.450548274037653e-05, |
| "loss": 9.145, |
| "num_input_tokens_seen": 100663296, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.06618021381299848, |
| "grad_norm": 3.035719394683838, |
| "learning_rate": 9.420211310767533e-05, |
| "loss": 8.8996, |
| "num_input_tokens_seen": 102236160, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.06719837094858307, |
| "grad_norm": 2.5634443759918213, |
| "learning_rate": 9.389110615965102e-05, |
| "loss": 8.8526, |
| "num_input_tokens_seen": 103809024, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.06821652808416766, |
| "grad_norm": 2.726071834564209, |
| "learning_rate": 9.35725156326063e-05, |
| "loss": 8.6106, |
| "num_input_tokens_seen": 105381888, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.06923468521975225, |
| "grad_norm": 7.38182258605957, |
| "learning_rate": 9.324639657314742e-05, |
| "loss": 9.3592, |
| "num_input_tokens_seen": 106954752, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.07025284235533684, |
| "grad_norm": 2.8825085163116455, |
| "learning_rate": 9.291280532867302e-05, |
| "loss": 9.407, |
| "num_input_tokens_seen": 108527616, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.07127099949092143, |
| "grad_norm": 3.041630506515503, |
| "learning_rate": 9.257179953763845e-05, |
| "loss": 9.2032, |
| "num_input_tokens_seen": 110100480, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.07228915662650602, |
| "grad_norm": 2.5680899620056152, |
| "learning_rate": 9.222343811959693e-05, |
| "loss": 10.1544, |
| "num_input_tokens_seen": 111673344, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.07330731376209061, |
| "grad_norm": 3.0244016647338867, |
| "learning_rate": 9.186778126501916e-05, |
| "loss": 8.8151, |
| "num_input_tokens_seen": 113246208, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.0743254708976752, |
| "grad_norm": 3.2452633380889893, |
| "learning_rate": 9.150489042489367e-05, |
| "loss": 8.2984, |
| "num_input_tokens_seen": 114819072, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.0753436280332598, |
| "grad_norm": 3.0332398414611816, |
| "learning_rate": 9.113482830010918e-05, |
| "loss": 8.9962, |
| "num_input_tokens_seen": 116391936, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.0763617851688444, |
| "grad_norm": 2.4632184505462646, |
| "learning_rate": 9.075765883062093e-05, |
| "loss": 10.0713, |
| "num_input_tokens_seen": 117964800, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.07737994230442899, |
| "grad_norm": 2.421849012374878, |
| "learning_rate": 9.037344718440322e-05, |
| "loss": 9.5697, |
| "num_input_tokens_seen": 119537664, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.07839809944001358, |
| "grad_norm": 2.8447134494781494, |
| "learning_rate": 8.99822597461894e-05, |
| "loss": 10.0348, |
| "num_input_tokens_seen": 121110528, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.07941625657559817, |
| "grad_norm": 3.023327350616455, |
| "learning_rate": 8.958416410600187e-05, |
| "loss": 8.2245, |
| "num_input_tokens_seen": 122683392, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.08043441371118276, |
| "grad_norm": 2.6515488624572754, |
| "learning_rate": 8.917922904747384e-05, |
| "loss": 9.5885, |
| "num_input_tokens_seen": 124256256, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.08145257084676735, |
| "grad_norm": 2.6909425258636475, |
| "learning_rate": 8.876752453596462e-05, |
| "loss": 7.7732, |
| "num_input_tokens_seen": 125829120, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.08247072798235194, |
| "grad_norm": 2.8426523208618164, |
| "learning_rate": 8.834912170647101e-05, |
| "loss": 9.0941, |
| "num_input_tokens_seen": 127401984, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.08348888511793653, |
| "grad_norm": 23.567190170288086, |
| "learning_rate": 8.792409285133642e-05, |
| "loss": 9.6, |
| "num_input_tokens_seen": 128974848, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.08450704225352113, |
| "grad_norm": 2.7843549251556396, |
| "learning_rate": 8.749251140776016e-05, |
| "loss": 7.8768, |
| "num_input_tokens_seen": 130547712, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.08552519938910572, |
| "grad_norm": 2.600492477416992, |
| "learning_rate": 8.705445194510868e-05, |
| "loss": 9.2542, |
| "num_input_tokens_seen": 132120576, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.0865433565246903, |
| "grad_norm": 2.8346362113952637, |
| "learning_rate": 8.66099901520315e-05, |
| "loss": 9.481, |
| "num_input_tokens_seen": 133693440, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.0875615136602749, |
| "grad_norm": 2.405304193496704, |
| "learning_rate": 8.615920282338355e-05, |
| "loss": 8.7034, |
| "num_input_tokens_seen": 135266304, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.08857967079585949, |
| "grad_norm": 3.806959390640259, |
| "learning_rate": 8.570216784695637e-05, |
| "loss": 8.5569, |
| "num_input_tokens_seen": 136839168, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.08959782793144408, |
| "grad_norm": 3.082869052886963, |
| "learning_rate": 8.52389641900206e-05, |
| "loss": 9.3447, |
| "num_input_tokens_seen": 138412032, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.09061598506702868, |
| "grad_norm": 3.056046724319458, |
| "learning_rate": 8.476967188568188e-05, |
| "loss": 8.0853, |
| "num_input_tokens_seen": 139984896, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.09163414220261328, |
| "grad_norm": 2.4092164039611816, |
| "learning_rate": 8.429437201905254e-05, |
| "loss": 9.6728, |
| "num_input_tokens_seen": 141557760, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.09265229933819787, |
| "grad_norm": 2.415492057800293, |
| "learning_rate": 8.381314671324159e-05, |
| "loss": 9.8961, |
| "num_input_tokens_seen": 143130624, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.09367045647378246, |
| "grad_norm": 2.674098253250122, |
| "learning_rate": 8.332607911516545e-05, |
| "loss": 9.4862, |
| "num_input_tokens_seen": 144703488, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.09468861360936705, |
| "grad_norm": 3.606937885284424, |
| "learning_rate": 8.283325338118153e-05, |
| "loss": 7.985, |
| "num_input_tokens_seen": 146276352, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.09570677074495164, |
| "grad_norm": 2.8389623165130615, |
| "learning_rate": 8.233475466254765e-05, |
| "loss": 8.734, |
| "num_input_tokens_seen": 147849216, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.09672492788053623, |
| "grad_norm": 25.968826293945312, |
| "learning_rate": 8.183066909070947e-05, |
| "loss": 9.4758, |
| "num_input_tokens_seen": 149422080, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.09774308501612082, |
| "grad_norm": 3.1141164302825928, |
| "learning_rate": 8.132108376241849e-05, |
| "loss": 7.9685, |
| "num_input_tokens_seen": 150994944, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.09876124215170541, |
| "grad_norm": 2.62211275100708, |
| "learning_rate": 8.08060867246834e-05, |
| "loss": 8.5738, |
| "num_input_tokens_seen": 152567808, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.09977939928729, |
| "grad_norm": 3.2089600563049316, |
| "learning_rate": 8.028576695955711e-05, |
| "loss": 8.872, |
| "num_input_tokens_seen": 154140672, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.10079755642287459, |
| "grad_norm": 2.3852243423461914, |
| "learning_rate": 7.97602143687623e-05, |
| "loss": 8.9487, |
| "num_input_tokens_seen": 155713536, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.10181571355845918, |
| "grad_norm": 1.760627269744873, |
| "learning_rate": 7.922951975815811e-05, |
| "loss": 8.0331, |
| "num_input_tokens_seen": 157286400, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.10283387069404377, |
| "grad_norm": 2.6850216388702393, |
| "learning_rate": 7.869377482205042e-05, |
| "loss": 8.7944, |
| "num_input_tokens_seen": 158859264, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.10385202782962838, |
| "grad_norm": 2.7767252922058105, |
| "learning_rate": 7.815307212734888e-05, |
| "loss": 8.2323, |
| "num_input_tokens_seen": 160432128, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.10487018496521297, |
| "grad_norm": 2.437563896179199, |
| "learning_rate": 7.760750509757298e-05, |
| "loss": 9.859, |
| "num_input_tokens_seen": 162004992, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.10588834210079756, |
| "grad_norm": 2.759857177734375, |
| "learning_rate": 7.705716799671019e-05, |
| "loss": 8.3508, |
| "num_input_tokens_seen": 163577856, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.10690649923638215, |
| "grad_norm": 2.97436785697937, |
| "learning_rate": 7.650215591292888e-05, |
| "loss": 8.4575, |
| "num_input_tokens_seen": 165150720, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.10792465637196674, |
| "grad_norm": 2.894146203994751, |
| "learning_rate": 7.594256474214882e-05, |
| "loss": 9.5371, |
| "num_input_tokens_seen": 166723584, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.10894281350755133, |
| "grad_norm": 2.664741039276123, |
| "learning_rate": 7.537849117147212e-05, |
| "loss": 8.1767, |
| "num_input_tokens_seen": 168296448, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.10996097064313592, |
| "grad_norm": 2.5504794120788574, |
| "learning_rate": 7.481003266247744e-05, |
| "loss": 8.7513, |
| "num_input_tokens_seen": 169869312, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.11097912777872052, |
| "grad_norm": 2.3367724418640137, |
| "learning_rate": 7.423728743438048e-05, |
| "loss": 9.5805, |
| "num_input_tokens_seen": 171442176, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.1119972849143051, |
| "grad_norm": 2.3378474712371826, |
| "learning_rate": 7.366035444706347e-05, |
| "loss": 9.8007, |
| "num_input_tokens_seen": 173015040, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.1130154420498897, |
| "grad_norm": 3.5393240451812744, |
| "learning_rate": 7.307933338397667e-05, |
| "loss": 9.0731, |
| "num_input_tokens_seen": 174587904, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.11403359918547429, |
| "grad_norm": 2.569704532623291, |
| "learning_rate": 7.249432463491498e-05, |
| "loss": 8.8012, |
| "num_input_tokens_seen": 176160768, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.11505175632105888, |
| "grad_norm": 2.5539438724517822, |
| "learning_rate": 7.190542927867234e-05, |
| "loss": 8.9157, |
| "num_input_tokens_seen": 177733632, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.11606991345664347, |
| "grad_norm": 3.5465245246887207, |
| "learning_rate": 7.131274906557725e-05, |
| "loss": 8.9368, |
| "num_input_tokens_seen": 179306496, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.11708807059222806, |
| "grad_norm": 2.584242105484009, |
| "learning_rate": 7.071638639991207e-05, |
| "loss": 9.3932, |
| "num_input_tokens_seen": 180879360, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.11810622772781267, |
| "grad_norm": 2.462211847305298, |
| "learning_rate": 7.011644432221958e-05, |
| "loss": 9.9608, |
| "num_input_tokens_seen": 182452224, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.11912438486339726, |
| "grad_norm": 2.8939926624298096, |
| "learning_rate": 6.95130264914993e-05, |
| "loss": 8.8036, |
| "num_input_tokens_seen": 184025088, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.12014254199898185, |
| "grad_norm": 2.7085700035095215, |
| "learning_rate": 6.890623716729724e-05, |
| "loss": 9.6046, |
| "num_input_tokens_seen": 185597952, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.12116069913456644, |
| "grad_norm": 2.4249041080474854, |
| "learning_rate": 6.82961811916917e-05, |
| "loss": 8.4684, |
| "num_input_tokens_seen": 187170816, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.12217885627015103, |
| "grad_norm": 2.345521926879883, |
| "learning_rate": 6.768296397117848e-05, |
| "loss": 8.4042, |
| "num_input_tokens_seen": 188743680, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.12319701340573562, |
| "grad_norm": 4.278283596038818, |
| "learning_rate": 6.706669145845863e-05, |
| "loss": 9.243, |
| "num_input_tokens_seen": 190316544, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.12421517054132021, |
| "grad_norm": 3.101922035217285, |
| "learning_rate": 6.644747013413168e-05, |
| "loss": 8.6733, |
| "num_input_tokens_seen": 191889408, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.12523332767690482, |
| "grad_norm": 3.8298826217651367, |
| "learning_rate": 6.582540698829781e-05, |
| "loss": 9.3599, |
| "num_input_tokens_seen": 193462272, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.1262514848124894, |
| "grad_norm": 2.876701831817627, |
| "learning_rate": 6.520060950207185e-05, |
| "loss": 9.2528, |
| "num_input_tokens_seen": 195035136, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.127269641948074, |
| "grad_norm": 2.6711254119873047, |
| "learning_rate": 6.457318562901256e-05, |
| "loss": 7.5999, |
| "num_input_tokens_seen": 196608000, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.1282877990836586, |
| "grad_norm": 2.2993829250335693, |
| "learning_rate": 6.394324377647028e-05, |
| "loss": 9.1343, |
| "num_input_tokens_seen": 198180864, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.12930595621924318, |
| "grad_norm": 3.1544792652130127, |
| "learning_rate": 6.331089278685599e-05, |
| "loss": 8.7503, |
| "num_input_tokens_seen": 199753728, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.13032411335482777, |
| "grad_norm": 2.3709182739257812, |
| "learning_rate": 6.26762419188355e-05, |
| "loss": 8.1339, |
| "num_input_tokens_seen": 201326592, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.13134227049041236, |
| "grad_norm": 2.0676522254943848, |
| "learning_rate": 6.203940082845144e-05, |
| "loss": 8.6629, |
| "num_input_tokens_seen": 202899456, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.13236042762599695, |
| "grad_norm": 3.4315857887268066, |
| "learning_rate": 6.140047955017671e-05, |
| "loss": 8.5242, |
| "num_input_tokens_seen": 204472320, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.13337858476158154, |
| "grad_norm": 2.058745861053467, |
| "learning_rate": 6.075958847790262e-05, |
| "loss": 8.8911, |
| "num_input_tokens_seen": 206045184, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.13439674189716613, |
| "grad_norm": 2.168302297592163, |
| "learning_rate": 6.011683834586473e-05, |
| "loss": 8.767, |
| "num_input_tokens_seen": 207618048, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.13541489903275072, |
| "grad_norm": 2.312222480773926, |
| "learning_rate": 5.947234020951015e-05, |
| "loss": 8.8736, |
| "num_input_tokens_seen": 209190912, |
| "step": 133 |
| } |
| ], |
| "logging_steps": 1.0, |
| "max_steps": 266, |
| "num_input_tokens_seen": 209190912, |
| "num_train_epochs": 1, |
| "save_steps": 133, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 333751171153920.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|