| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.1359452956130453, |
| "eval_steps": 500, |
| "global_step": 10000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0013594529561304532, |
| "grad_norm": 2.6843507289886475, |
| "learning_rate": 2.9959624247202926e-05, |
| "loss": 0.791, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.0027189059122609063, |
| "grad_norm": 7.921170711517334, |
| "learning_rate": 2.9918840658519014e-05, |
| "loss": 0.7116, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.0040783588683913595, |
| "grad_norm": 1.703231692314148, |
| "learning_rate": 2.98780570698351e-05, |
| "loss": 0.7025, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.005437811824521813, |
| "grad_norm": 1.4689126014709473, |
| "learning_rate": 2.9837273481151187e-05, |
| "loss": 0.7026, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.006797264780652266, |
| "grad_norm": 6.632850646972656, |
| "learning_rate": 2.979648989246727e-05, |
| "loss": 0.7006, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.008156717736782719, |
| "grad_norm": 1.7937493324279785, |
| "learning_rate": 2.975570630378336e-05, |
| "loss": 0.6996, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.009516170692913172, |
| "grad_norm": 1.545050024986267, |
| "learning_rate": 2.9714922715099447e-05, |
| "loss": 0.6988, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.010875623649043625, |
| "grad_norm": 6.718420505523682, |
| "learning_rate": 2.967413912641553e-05, |
| "loss": 0.6934, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.012235076605174078, |
| "grad_norm": 2.417815923690796, |
| "learning_rate": 2.9633355537731616e-05, |
| "loss": 0.7031, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.013594529561304532, |
| "grad_norm": 3.016540765762329, |
| "learning_rate": 2.9592571949047704e-05, |
| "loss": 0.7003, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.014953982517434985, |
| "grad_norm": 1.3102728128433228, |
| "learning_rate": 2.9551788360363792e-05, |
| "loss": 0.6939, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.016313435473565438, |
| "grad_norm": 3.511146306991577, |
| "learning_rate": 2.9511004771679877e-05, |
| "loss": 0.7018, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.01767288842969589, |
| "grad_norm": 2.3049850463867188, |
| "learning_rate": 2.9470221182995965e-05, |
| "loss": 0.7093, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.019032341385826344, |
| "grad_norm": 1.0154485702514648, |
| "learning_rate": 2.942943759431205e-05, |
| "loss": 0.7035, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.020391794341956797, |
| "grad_norm": 2.085158348083496, |
| "learning_rate": 2.9388654005628134e-05, |
| "loss": 0.7015, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.02175124729808725, |
| "grad_norm": 3.9277114868164062, |
| "learning_rate": 2.9347870416944225e-05, |
| "loss": 0.6994, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.023110700254217704, |
| "grad_norm": 1.8050942420959473, |
| "learning_rate": 2.930708682826031e-05, |
| "loss": 0.7027, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.024470153210348157, |
| "grad_norm": 1.5299062728881836, |
| "learning_rate": 2.9266303239576394e-05, |
| "loss": 0.6989, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.02582960616647861, |
| "grad_norm": 3.9614956378936768, |
| "learning_rate": 2.9225519650892482e-05, |
| "loss": 0.6968, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.027189059122609063, |
| "grad_norm": 1.5142548084259033, |
| "learning_rate": 2.9184736062208567e-05, |
| "loss": 0.6996, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.028548512078739516, |
| "grad_norm": 2.0255579948425293, |
| "learning_rate": 2.9143952473524655e-05, |
| "loss": 0.6992, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.02990796503486997, |
| "grad_norm": 3.3495869636535645, |
| "learning_rate": 2.9103168884840743e-05, |
| "loss": 0.6986, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.03126741799100042, |
| "grad_norm": 4.41119909286499, |
| "learning_rate": 2.9062385296156827e-05, |
| "loss": 0.6963, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.032626870947130876, |
| "grad_norm": 2.923621416091919, |
| "learning_rate": 2.902160170747291e-05, |
| "loss": 0.6935, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.033986323903261326, |
| "grad_norm": 6.4013285636901855, |
| "learning_rate": 2.8980818118789e-05, |
| "loss": 0.6984, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.03534577685939178, |
| "grad_norm": 2.191399097442627, |
| "learning_rate": 2.8940034530105088e-05, |
| "loss": 0.6955, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.03670522981552223, |
| "grad_norm": 1.6993489265441895, |
| "learning_rate": 2.8899250941421172e-05, |
| "loss": 0.7002, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.03806468277165269, |
| "grad_norm": 3.364286184310913, |
| "learning_rate": 2.885846735273726e-05, |
| "loss": 0.6964, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.03942413572778314, |
| "grad_norm": 1.9982099533081055, |
| "learning_rate": 2.8817683764053345e-05, |
| "loss": 0.7069, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.040783588683913595, |
| "grad_norm": 1.524899959564209, |
| "learning_rate": 2.877690017536943e-05, |
| "loss": 0.6961, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.042143041640044045, |
| "grad_norm": 2.2762181758880615, |
| "learning_rate": 2.873611658668552e-05, |
| "loss": 0.6991, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.0435024945961745, |
| "grad_norm": 2.0052831172943115, |
| "learning_rate": 2.8695332998001605e-05, |
| "loss": 0.6987, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.04486194755230495, |
| "grad_norm": 1.6109389066696167, |
| "learning_rate": 2.865454940931769e-05, |
| "loss": 0.6974, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.04622140050843541, |
| "grad_norm": 1.7458642721176147, |
| "learning_rate": 2.8613765820633778e-05, |
| "loss": 0.6984, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.04758085346456586, |
| "grad_norm": 5.7916083335876465, |
| "learning_rate": 2.8572982231949862e-05, |
| "loss": 0.6986, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.048940306420696314, |
| "grad_norm": 2.8352317810058594, |
| "learning_rate": 2.8532198643265954e-05, |
| "loss": 0.7004, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.050299759376826764, |
| "grad_norm": 6.102287769317627, |
| "learning_rate": 2.8491415054582038e-05, |
| "loss": 0.6975, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.05165921233295722, |
| "grad_norm": 2.8221065998077393, |
| "learning_rate": 2.8450631465898123e-05, |
| "loss": 0.7092, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.05301866528908767, |
| "grad_norm": 2.0802714824676514, |
| "learning_rate": 2.840984787721421e-05, |
| "loss": 0.7097, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.05437811824521813, |
| "grad_norm": 4.515227794647217, |
| "learning_rate": 2.8369064288530295e-05, |
| "loss": 0.6991, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.055737571201348576, |
| "grad_norm": 4.737243175506592, |
| "learning_rate": 2.8328280699846383e-05, |
| "loss": 0.7042, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.05709702415747903, |
| "grad_norm": 5.016382694244385, |
| "learning_rate": 2.828749711116247e-05, |
| "loss": 0.7008, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.05845647711360948, |
| "grad_norm": 2.7722420692443848, |
| "learning_rate": 2.8246713522478556e-05, |
| "loss": 0.7038, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.05981593006973994, |
| "grad_norm": 3.596283435821533, |
| "learning_rate": 2.820592993379464e-05, |
| "loss": 0.7008, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.06117538302587039, |
| "grad_norm": 3.8389534950256348, |
| "learning_rate": 2.8165146345110728e-05, |
| "loss": 0.6986, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.06253483598200084, |
| "grad_norm": 2.992204189300537, |
| "learning_rate": 2.8124362756426816e-05, |
| "loss": 0.7027, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.0638942889381313, |
| "grad_norm": 1.4091521501541138, |
| "learning_rate": 2.80835791677429e-05, |
| "loss": 0.7032, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.06525374189426175, |
| "grad_norm": 1.7205729484558105, |
| "learning_rate": 2.804279557905899e-05, |
| "loss": 0.7061, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.0666131948503922, |
| "grad_norm": 2.666346549987793, |
| "learning_rate": 2.8002011990375073e-05, |
| "loss": 0.7016, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.06797264780652265, |
| "grad_norm": 2.0234336853027344, |
| "learning_rate": 2.7961228401691158e-05, |
| "loss": 0.6986, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.06933210076265311, |
| "grad_norm": 1.3103491067886353, |
| "learning_rate": 2.792044481300725e-05, |
| "loss": 0.6995, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.07069155371878356, |
| "grad_norm": 3.3153600692749023, |
| "learning_rate": 2.7879661224323334e-05, |
| "loss": 0.6968, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.07205100667491401, |
| "grad_norm": 3.262269973754883, |
| "learning_rate": 2.7838877635639418e-05, |
| "loss": 0.7041, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.07341045963104446, |
| "grad_norm": 5.839659214019775, |
| "learning_rate": 2.7798094046955506e-05, |
| "loss": 0.692, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.07476991258717493, |
| "grad_norm": 6.034287929534912, |
| "learning_rate": 2.775731045827159e-05, |
| "loss": 0.7034, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.07612936554330538, |
| "grad_norm": 3.7333922386169434, |
| "learning_rate": 2.771652686958768e-05, |
| "loss": 0.6986, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.07748881849943583, |
| "grad_norm": 2.81872820854187, |
| "learning_rate": 2.7675743280903767e-05, |
| "loss": 0.7021, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.07884827145556628, |
| "grad_norm": 2.6562986373901367, |
| "learning_rate": 2.763495969221985e-05, |
| "loss": 0.6999, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.08020772441169673, |
| "grad_norm": 4.191847801208496, |
| "learning_rate": 2.7594176103535936e-05, |
| "loss": 0.704, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.08156717736782719, |
| "grad_norm": 1.9238234758377075, |
| "learning_rate": 2.7553392514852024e-05, |
| "loss": 0.6996, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.08292663032395764, |
| "grad_norm": 1.6448299884796143, |
| "learning_rate": 2.751260892616811e-05, |
| "loss": 0.7033, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.08428608328008809, |
| "grad_norm": 2.8520469665527344, |
| "learning_rate": 2.7471825337484196e-05, |
| "loss": 0.707, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.08564553623621854, |
| "grad_norm": 4.698349952697754, |
| "learning_rate": 2.7431041748800284e-05, |
| "loss": 0.6995, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.087004989192349, |
| "grad_norm": 3.2636826038360596, |
| "learning_rate": 2.739025816011637e-05, |
| "loss": 0.697, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.08836444214847945, |
| "grad_norm": 5.062309741973877, |
| "learning_rate": 2.7349474571432457e-05, |
| "loss": 0.7068, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.0897238951046099, |
| "grad_norm": 1.9477702379226685, |
| "learning_rate": 2.7308690982748545e-05, |
| "loss": 0.6946, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.09108334806074035, |
| "grad_norm": 6.437952518463135, |
| "learning_rate": 2.726790739406463e-05, |
| "loss": 0.6968, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.09244280101687082, |
| "grad_norm": 1.488918423652649, |
| "learning_rate": 2.7227123805380717e-05, |
| "loss": 0.6964, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.09380225397300126, |
| "grad_norm": 1.5171183347702026, |
| "learning_rate": 2.7186340216696802e-05, |
| "loss": 0.6948, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.09516170692913171, |
| "grad_norm": 5.248293876647949, |
| "learning_rate": 2.7145556628012886e-05, |
| "loss": 0.6914, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.09652115988526216, |
| "grad_norm": 3.5038247108459473, |
| "learning_rate": 2.7104773039328978e-05, |
| "loss": 0.7016, |
| "step": 7100 |
| }, |
| { |
| "epoch": 0.09788061284139263, |
| "grad_norm": 2.2439801692962646, |
| "learning_rate": 2.7063989450645062e-05, |
| "loss": 0.6976, |
| "step": 7200 |
| }, |
| { |
| "epoch": 0.09924006579752308, |
| "grad_norm": 5.262351036071777, |
| "learning_rate": 2.7023205861961147e-05, |
| "loss": 0.7015, |
| "step": 7300 |
| }, |
| { |
| "epoch": 0.10059951875365353, |
| "grad_norm": 2.734067916870117, |
| "learning_rate": 2.6982422273277235e-05, |
| "loss": 0.7012, |
| "step": 7400 |
| }, |
| { |
| "epoch": 0.10195897170978398, |
| "grad_norm": 7.341092586517334, |
| "learning_rate": 2.694163868459332e-05, |
| "loss": 0.6977, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.10331842466591444, |
| "grad_norm": 2.047778367996216, |
| "learning_rate": 2.6900855095909407e-05, |
| "loss": 0.6934, |
| "step": 7600 |
| }, |
| { |
| "epoch": 0.10467787762204489, |
| "grad_norm": 5.612318515777588, |
| "learning_rate": 2.6860071507225495e-05, |
| "loss": 0.696, |
| "step": 7700 |
| }, |
| { |
| "epoch": 0.10603733057817534, |
| "grad_norm": 3.8864567279815674, |
| "learning_rate": 2.681928791854158e-05, |
| "loss": 0.6958, |
| "step": 7800 |
| }, |
| { |
| "epoch": 0.10739678353430579, |
| "grad_norm": 2.456672191619873, |
| "learning_rate": 2.6778504329857664e-05, |
| "loss": 0.7014, |
| "step": 7900 |
| }, |
| { |
| "epoch": 0.10875623649043625, |
| "grad_norm": 1.5562827587127686, |
| "learning_rate": 2.6737720741173752e-05, |
| "loss": 0.7, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.1101156894465667, |
| "grad_norm": 1.646262764930725, |
| "learning_rate": 2.669693715248984e-05, |
| "loss": 0.7028, |
| "step": 8100 |
| }, |
| { |
| "epoch": 0.11147514240269715, |
| "grad_norm": 2.5881056785583496, |
| "learning_rate": 2.6656153563805925e-05, |
| "loss": 0.697, |
| "step": 8200 |
| }, |
| { |
| "epoch": 0.1128345953588276, |
| "grad_norm": 3.108797788619995, |
| "learning_rate": 2.6615369975122013e-05, |
| "loss": 0.7, |
| "step": 8300 |
| }, |
| { |
| "epoch": 0.11419404831495807, |
| "grad_norm": 1.5215388536453247, |
| "learning_rate": 2.6574586386438097e-05, |
| "loss": 0.6956, |
| "step": 8400 |
| }, |
| { |
| "epoch": 0.11555350127108852, |
| "grad_norm": 1.2557023763656616, |
| "learning_rate": 2.6533802797754182e-05, |
| "loss": 0.7032, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.11691295422721897, |
| "grad_norm": 3.6592652797698975, |
| "learning_rate": 2.6493019209070273e-05, |
| "loss": 0.6957, |
| "step": 8600 |
| }, |
| { |
| "epoch": 0.11827240718334942, |
| "grad_norm": 1.6143642663955688, |
| "learning_rate": 2.6452235620386358e-05, |
| "loss": 0.6961, |
| "step": 8700 |
| }, |
| { |
| "epoch": 0.11963186013947988, |
| "grad_norm": 1.7212355136871338, |
| "learning_rate": 2.6411452031702442e-05, |
| "loss": 0.7041, |
| "step": 8800 |
| }, |
| { |
| "epoch": 0.12099131309561033, |
| "grad_norm": 1.2407207489013672, |
| "learning_rate": 2.637066844301853e-05, |
| "loss": 0.6938, |
| "step": 8900 |
| }, |
| { |
| "epoch": 0.12235076605174078, |
| "grad_norm": 4.306702613830566, |
| "learning_rate": 2.6329884854334615e-05, |
| "loss": 0.704, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.12371021900787123, |
| "grad_norm": 4.0667219161987305, |
| "learning_rate": 2.6289101265650703e-05, |
| "loss": 0.6966, |
| "step": 9100 |
| }, |
| { |
| "epoch": 0.12506967196400168, |
| "grad_norm": 2.244699478149414, |
| "learning_rate": 2.624831767696679e-05, |
| "loss": 0.6976, |
| "step": 9200 |
| }, |
| { |
| "epoch": 0.12642912492013214, |
| "grad_norm": 4.839937210083008, |
| "learning_rate": 2.6207534088282875e-05, |
| "loss": 0.6972, |
| "step": 9300 |
| }, |
| { |
| "epoch": 0.1277885778762626, |
| "grad_norm": 5.436954021453857, |
| "learning_rate": 2.6166750499598963e-05, |
| "loss": 0.698, |
| "step": 9400 |
| }, |
| { |
| "epoch": 0.12914803083239304, |
| "grad_norm": 5.324636459350586, |
| "learning_rate": 2.6125966910915048e-05, |
| "loss": 0.699, |
| "step": 9500 |
| }, |
| { |
| "epoch": 0.1305074837885235, |
| "grad_norm": 2.4143927097320557, |
| "learning_rate": 2.6085183322231136e-05, |
| "loss": 0.7033, |
| "step": 9600 |
| }, |
| { |
| "epoch": 0.13186693674465394, |
| "grad_norm": 3.336245059967041, |
| "learning_rate": 2.6044399733547224e-05, |
| "loss": 0.6993, |
| "step": 9700 |
| }, |
| { |
| "epoch": 0.1332263897007844, |
| "grad_norm": 1.2645655870437622, |
| "learning_rate": 2.600361614486331e-05, |
| "loss": 0.6977, |
| "step": 9800 |
| }, |
| { |
| "epoch": 0.13458584265691487, |
| "grad_norm": 3.2803938388824463, |
| "learning_rate": 2.5962832556179393e-05, |
| "loss": 0.6993, |
| "step": 9900 |
| }, |
| { |
| "epoch": 0.1359452956130453, |
| "grad_norm": 2.2295751571655273, |
| "learning_rate": 2.592204896749548e-05, |
| "loss": 0.6923, |
| "step": 10000 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 73559, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 10000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|