| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.08187912594033059, | |
| "eval_steps": 500, | |
| "global_step": 400, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.00020469781485082646, | |
| "grad_norm": 1.694807571622714, | |
| "learning_rate": 1.3605442176870747e-08, | |
| "loss": 0.6346, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0004093956297016529, | |
| "grad_norm": 1.7333604659657242, | |
| "learning_rate": 2.7210884353741493e-08, | |
| "loss": 0.5956, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0006140934445524794, | |
| "grad_norm": 1.675170156463719, | |
| "learning_rate": 4.081632653061224e-08, | |
| "loss": 0.6494, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.0008187912594033058, | |
| "grad_norm": 1.4535412459526658, | |
| "learning_rate": 5.442176870748299e-08, | |
| "loss": 0.6097, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.0010234890742541324, | |
| "grad_norm": 1.6204240919715567, | |
| "learning_rate": 6.802721088435375e-08, | |
| "loss": 0.6259, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0012281868891049587, | |
| "grad_norm": 1.6382909584778356, | |
| "learning_rate": 8.163265306122448e-08, | |
| "loss": 0.6049, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.0014328847039557853, | |
| "grad_norm": 1.6591863389965569, | |
| "learning_rate": 9.523809523809523e-08, | |
| "loss": 0.6093, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.0016375825188066117, | |
| "grad_norm": 1.529188807208944, | |
| "learning_rate": 1.0884353741496597e-07, | |
| "loss": 0.625, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.0018422803336574382, | |
| "grad_norm": 1.7414059653199376, | |
| "learning_rate": 1.2244897959183673e-07, | |
| "loss": 0.6148, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.002046978148508265, | |
| "grad_norm": 1.6622320550472127, | |
| "learning_rate": 1.360544217687075e-07, | |
| "loss": 0.5797, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.002251675963359091, | |
| "grad_norm": 1.6508189144245708, | |
| "learning_rate": 1.4965986394557823e-07, | |
| "loss": 0.6484, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.0024563737782099175, | |
| "grad_norm": 1.7202133207821506, | |
| "learning_rate": 1.6326530612244896e-07, | |
| "loss": 0.6216, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.0026610715930607443, | |
| "grad_norm": 1.5235060143030161, | |
| "learning_rate": 1.7687074829931972e-07, | |
| "loss": 0.6452, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.0028657694079115706, | |
| "grad_norm": 1.5363560852946705, | |
| "learning_rate": 1.9047619047619045e-07, | |
| "loss": 0.6405, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.003070467222762397, | |
| "grad_norm": 1.6730999257251689, | |
| "learning_rate": 2.0408163265306121e-07, | |
| "loss": 0.6497, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.0032751650376132233, | |
| "grad_norm": 1.5758327138243107, | |
| "learning_rate": 2.1768707482993195e-07, | |
| "loss": 0.6336, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.00347986285246405, | |
| "grad_norm": 1.5492535238923828, | |
| "learning_rate": 2.312925170068027e-07, | |
| "loss": 0.6037, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.0036845606673148765, | |
| "grad_norm": 1.6696926699572276, | |
| "learning_rate": 2.4489795918367347e-07, | |
| "loss": 0.6139, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.003889258482165703, | |
| "grad_norm": 1.6544769292475, | |
| "learning_rate": 2.5850340136054423e-07, | |
| "loss": 0.6315, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.00409395629701653, | |
| "grad_norm": 1.6860896587110352, | |
| "learning_rate": 2.72108843537415e-07, | |
| "loss": 0.6324, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0042986541118673555, | |
| "grad_norm": 1.5451565683271684, | |
| "learning_rate": 2.857142857142857e-07, | |
| "loss": 0.6227, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.004503351926718182, | |
| "grad_norm": 1.5658617650258626, | |
| "learning_rate": 2.9931972789115645e-07, | |
| "loss": 0.5873, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.004708049741569009, | |
| "grad_norm": 1.6014268573351107, | |
| "learning_rate": 3.129251700680272e-07, | |
| "loss": 0.6066, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.004912747556419835, | |
| "grad_norm": 1.560124972985785, | |
| "learning_rate": 3.265306122448979e-07, | |
| "loss": 0.6062, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.005117445371270662, | |
| "grad_norm": 1.6662191020723245, | |
| "learning_rate": 3.401360544217687e-07, | |
| "loss": 0.5968, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.0053221431861214885, | |
| "grad_norm": 1.7137595321931511, | |
| "learning_rate": 3.5374149659863944e-07, | |
| "loss": 0.6325, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.0055268410009723145, | |
| "grad_norm": 1.570188696390546, | |
| "learning_rate": 3.673469387755102e-07, | |
| "loss": 0.6375, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.005731538815823141, | |
| "grad_norm": 1.5585726347237283, | |
| "learning_rate": 3.809523809523809e-07, | |
| "loss": 0.6216, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.005936236630673967, | |
| "grad_norm": 1.4585947364133294, | |
| "learning_rate": 3.9455782312925167e-07, | |
| "loss": 0.5975, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.006140934445524794, | |
| "grad_norm": 1.5313204391085877, | |
| "learning_rate": 4.0816326530612243e-07, | |
| "loss": 0.6461, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.006345632260375621, | |
| "grad_norm": 1.4690318982818216, | |
| "learning_rate": 4.217687074829932e-07, | |
| "loss": 0.6054, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.006550330075226447, | |
| "grad_norm": 1.5256726957060316, | |
| "learning_rate": 4.353741496598639e-07, | |
| "loss": 0.6507, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.006755027890077273, | |
| "grad_norm": 1.541131533646238, | |
| "learning_rate": 4.4897959183673465e-07, | |
| "loss": 0.6185, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.0069597257049281, | |
| "grad_norm": 1.5233070330699345, | |
| "learning_rate": 4.625850340136054e-07, | |
| "loss": 0.6541, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.007164423519778926, | |
| "grad_norm": 1.4300240195672376, | |
| "learning_rate": 4.761904761904761e-07, | |
| "loss": 0.6156, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.007369121334629753, | |
| "grad_norm": 1.3386118655838508, | |
| "learning_rate": 4.897959183673469e-07, | |
| "loss": 0.5943, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.00757381914948058, | |
| "grad_norm": 1.3270928639031936, | |
| "learning_rate": 5.034013605442177e-07, | |
| "loss": 0.6139, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.007778516964331406, | |
| "grad_norm": 1.3116129547815811, | |
| "learning_rate": 5.170068027210885e-07, | |
| "loss": 0.6119, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.007983214779182232, | |
| "grad_norm": 1.2451803048665653, | |
| "learning_rate": 5.306122448979592e-07, | |
| "loss": 0.5463, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.00818791259403306, | |
| "grad_norm": 1.2351915311334578, | |
| "learning_rate": 5.4421768707483e-07, | |
| "loss": 0.5762, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.008392610408883885, | |
| "grad_norm": 1.3425104949855924, | |
| "learning_rate": 5.578231292517006e-07, | |
| "loss": 0.5866, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.008597308223734711, | |
| "grad_norm": 1.3464358053560985, | |
| "learning_rate": 5.714285714285714e-07, | |
| "loss": 0.6134, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.008802006038585539, | |
| "grad_norm": 1.3225968492677225, | |
| "learning_rate": 5.850340136054421e-07, | |
| "loss": 0.6034, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.009006703853436365, | |
| "grad_norm": 1.2483346937333237, | |
| "learning_rate": 5.986394557823129e-07, | |
| "loss": 0.5495, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.00921140166828719, | |
| "grad_norm": 1.1648688787665145, | |
| "learning_rate": 6.122448979591837e-07, | |
| "loss": 0.616, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.009416099483138018, | |
| "grad_norm": 1.2616996144445687, | |
| "learning_rate": 6.258503401360544e-07, | |
| "loss": 0.57, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.009620797297988844, | |
| "grad_norm": 1.3108653064941627, | |
| "learning_rate": 6.394557823129252e-07, | |
| "loss": 0.5814, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.00982549511283967, | |
| "grad_norm": 1.1754918916726766, | |
| "learning_rate": 6.530612244897958e-07, | |
| "loss": 0.5754, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.010030192927690498, | |
| "grad_norm": 1.272022559229399, | |
| "learning_rate": 6.666666666666666e-07, | |
| "loss": 0.5944, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.010234890742541324, | |
| "grad_norm": 1.13107848406085, | |
| "learning_rate": 6.802721088435374e-07, | |
| "loss": 0.5945, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.01043958855739215, | |
| "grad_norm": 1.1273813534766033, | |
| "learning_rate": 6.938775510204081e-07, | |
| "loss": 0.5538, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.010644286372242977, | |
| "grad_norm": 1.1293664677810216, | |
| "learning_rate": 7.074829931972789e-07, | |
| "loss": 0.5854, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.010848984187093803, | |
| "grad_norm": 0.9728651370750258, | |
| "learning_rate": 7.210884353741496e-07, | |
| "loss": 0.5108, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.011053682001944629, | |
| "grad_norm": 1.0432420839745669, | |
| "learning_rate": 7.346938775510204e-07, | |
| "loss": 0.5346, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.011258379816795457, | |
| "grad_norm": 1.0023551080535893, | |
| "learning_rate": 7.482993197278912e-07, | |
| "loss": 0.5799, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.011463077631646282, | |
| "grad_norm": 0.9638908320867696, | |
| "learning_rate": 7.619047619047618e-07, | |
| "loss": 0.555, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.011667775446497108, | |
| "grad_norm": 1.0398584356633989, | |
| "learning_rate": 7.755102040816326e-07, | |
| "loss": 0.5147, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.011872473261347934, | |
| "grad_norm": 0.9629896909635629, | |
| "learning_rate": 7.891156462585033e-07, | |
| "loss": 0.5413, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.012077171076198762, | |
| "grad_norm": 0.9770292637339174, | |
| "learning_rate": 8.027210884353741e-07, | |
| "loss": 0.5205, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.012281868891049588, | |
| "grad_norm": 0.971945782703798, | |
| "learning_rate": 8.163265306122449e-07, | |
| "loss": 0.5422, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.012486566705900414, | |
| "grad_norm": 0.950398975311517, | |
| "learning_rate": 8.299319727891156e-07, | |
| "loss": 0.5071, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.012691264520751241, | |
| "grad_norm": 0.9049285150490526, | |
| "learning_rate": 8.435374149659864e-07, | |
| "loss": 0.4964, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.012895962335602067, | |
| "grad_norm": 0.8793095995125478, | |
| "learning_rate": 8.57142857142857e-07, | |
| "loss": 0.5331, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.013100660150452893, | |
| "grad_norm": 0.8515461613654705, | |
| "learning_rate": 8.707482993197278e-07, | |
| "loss": 0.5283, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.013305357965303721, | |
| "grad_norm": 0.867859420385022, | |
| "learning_rate": 8.843537414965985e-07, | |
| "loss": 0.5164, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.013510055780154547, | |
| "grad_norm": 0.8786706131313361, | |
| "learning_rate": 8.979591836734693e-07, | |
| "loss": 0.5645, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.013714753595005373, | |
| "grad_norm": 0.8579092596142676, | |
| "learning_rate": 9.115646258503401e-07, | |
| "loss": 0.5399, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.0139194514098562, | |
| "grad_norm": 0.8773908463960428, | |
| "learning_rate": 9.251700680272108e-07, | |
| "loss": 0.5229, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.014124149224707026, | |
| "grad_norm": 0.8528366708567172, | |
| "learning_rate": 9.387755102040816e-07, | |
| "loss": 0.5349, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.014328847039557852, | |
| "grad_norm": 0.9184139371914097, | |
| "learning_rate": 9.523809523809522e-07, | |
| "loss": 0.5331, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.01453354485440868, | |
| "grad_norm": 0.8507461371837629, | |
| "learning_rate": 9.65986394557823e-07, | |
| "loss": 0.4801, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.014738242669259506, | |
| "grad_norm": 0.8374936253263676, | |
| "learning_rate": 9.795918367346939e-07, | |
| "loss": 0.4931, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.014942940484110332, | |
| "grad_norm": 0.8174848059151317, | |
| "learning_rate": 9.931972789115645e-07, | |
| "loss": 0.5248, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.01514763829896116, | |
| "grad_norm": 0.8174077531772923, | |
| "learning_rate": 1.0068027210884354e-06, | |
| "loss": 0.5036, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.015352336113811985, | |
| "grad_norm": 0.7262562022534738, | |
| "learning_rate": 1.020408163265306e-06, | |
| "loss": 0.5232, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.015557033928662811, | |
| "grad_norm": 0.7855250505927771, | |
| "learning_rate": 1.034013605442177e-06, | |
| "loss": 0.5098, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.015761731743513637, | |
| "grad_norm": 0.8278680336215173, | |
| "learning_rate": 1.0476190476190476e-06, | |
| "loss": 0.4829, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.015966429558364463, | |
| "grad_norm": 0.797196328457245, | |
| "learning_rate": 1.0612244897959184e-06, | |
| "loss": 0.5037, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.016171127373215292, | |
| "grad_norm": 0.7507210642711485, | |
| "learning_rate": 1.074829931972789e-06, | |
| "loss": 0.4944, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.01637582518806612, | |
| "grad_norm": 0.826047544790976, | |
| "learning_rate": 1.08843537414966e-06, | |
| "loss": 0.5179, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.016580523002916944, | |
| "grad_norm": 0.7746315656318813, | |
| "learning_rate": 1.1020408163265304e-06, | |
| "loss": 0.5223, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.01678522081776777, | |
| "grad_norm": 0.778762710130468, | |
| "learning_rate": 1.1156462585034013e-06, | |
| "loss": 0.4845, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.016989918632618596, | |
| "grad_norm": 0.749908717861716, | |
| "learning_rate": 1.129251700680272e-06, | |
| "loss": 0.5175, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.017194616447469422, | |
| "grad_norm": 0.7582554704845739, | |
| "learning_rate": 1.1428571428571428e-06, | |
| "loss": 0.4978, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.01739931426232025, | |
| "grad_norm": 0.7595367961287336, | |
| "learning_rate": 1.1564625850340134e-06, | |
| "loss": 0.4966, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.017604012077171077, | |
| "grad_norm": 0.7488555001974914, | |
| "learning_rate": 1.1700680272108843e-06, | |
| "loss": 0.5025, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.017808709892021903, | |
| "grad_norm": 0.8307772703305798, | |
| "learning_rate": 1.183673469387755e-06, | |
| "loss": 0.5144, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.01801340770687273, | |
| "grad_norm": 0.7317615547098743, | |
| "learning_rate": 1.1972789115646258e-06, | |
| "loss": 0.4817, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.018218105521723555, | |
| "grad_norm": 0.8210594860542216, | |
| "learning_rate": 1.2108843537414965e-06, | |
| "loss": 0.5058, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.01842280333657438, | |
| "grad_norm": 0.7250535412206353, | |
| "learning_rate": 1.2244897959183673e-06, | |
| "loss": 0.4796, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.018627501151425207, | |
| "grad_norm": 0.7476633557284366, | |
| "learning_rate": 1.238095238095238e-06, | |
| "loss": 0.4732, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.018832198966276036, | |
| "grad_norm": 0.7245302420505394, | |
| "learning_rate": 1.2517006802721089e-06, | |
| "loss": 0.5085, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.019036896781126862, | |
| "grad_norm": 0.7287781044325405, | |
| "learning_rate": 1.2653061224489795e-06, | |
| "loss": 0.4837, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.019241594595977688, | |
| "grad_norm": 0.7461257075758424, | |
| "learning_rate": 1.2789115646258504e-06, | |
| "loss": 0.4847, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.019446292410828514, | |
| "grad_norm": 0.7500567577642135, | |
| "learning_rate": 1.292517006802721e-06, | |
| "loss": 0.5023, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.01965099022567934, | |
| "grad_norm": 0.7516926737451503, | |
| "learning_rate": 1.3061224489795917e-06, | |
| "loss": 0.4944, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.019855688040530166, | |
| "grad_norm": 0.8160475040600308, | |
| "learning_rate": 1.3197278911564623e-06, | |
| "loss": 0.4707, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.020060385855380995, | |
| "grad_norm": 0.7313987935291313, | |
| "learning_rate": 1.3333333333333332e-06, | |
| "loss": 0.4631, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.02026508367023182, | |
| "grad_norm": 0.7272827048713341, | |
| "learning_rate": 1.3469387755102039e-06, | |
| "loss": 0.4912, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.020469781485082647, | |
| "grad_norm": 0.7148392974765637, | |
| "learning_rate": 1.3605442176870747e-06, | |
| "loss": 0.4686, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.020674479299933473, | |
| "grad_norm": 0.8073254642999934, | |
| "learning_rate": 1.3741496598639456e-06, | |
| "loss": 0.4889, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.0208791771147843, | |
| "grad_norm": 0.7585784341693678, | |
| "learning_rate": 1.3877551020408162e-06, | |
| "loss": 0.4661, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.021083874929635125, | |
| "grad_norm": 0.750059071249337, | |
| "learning_rate": 1.4013605442176871e-06, | |
| "loss": 0.4856, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.021288572744485954, | |
| "grad_norm": 0.7391246566572075, | |
| "learning_rate": 1.4149659863945578e-06, | |
| "loss": 0.4835, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.02149327055933678, | |
| "grad_norm": 0.7961401475792825, | |
| "learning_rate": 1.4285714285714286e-06, | |
| "loss": 0.5078, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.021697968374187606, | |
| "grad_norm": 0.736443177126423, | |
| "learning_rate": 1.4421768707482993e-06, | |
| "loss": 0.4754, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.021902666189038432, | |
| "grad_norm": 0.7433000385873849, | |
| "learning_rate": 1.4557823129251701e-06, | |
| "loss": 0.5051, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.022107364003889258, | |
| "grad_norm": 0.742852894387874, | |
| "learning_rate": 1.4693877551020408e-06, | |
| "loss": 0.4815, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.022312061818740084, | |
| "grad_norm": 0.7321778169129644, | |
| "learning_rate": 1.4829931972789117e-06, | |
| "loss": 0.4883, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.022516759633590913, | |
| "grad_norm": 0.7374200652655346, | |
| "learning_rate": 1.4965986394557823e-06, | |
| "loss": 0.5235, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.02272145744844174, | |
| "grad_norm": 0.772981855244519, | |
| "learning_rate": 1.510204081632653e-06, | |
| "loss": 0.5008, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.022926155263292565, | |
| "grad_norm": 0.7342674795579016, | |
| "learning_rate": 1.5238095238095236e-06, | |
| "loss": 0.4671, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.02313085307814339, | |
| "grad_norm": 0.7795902315585469, | |
| "learning_rate": 1.5374149659863945e-06, | |
| "loss": 0.507, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.023335550892994217, | |
| "grad_norm": 0.7765099211131105, | |
| "learning_rate": 1.5510204081632651e-06, | |
| "loss": 0.5251, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.023540248707845043, | |
| "grad_norm": 0.7386929957340117, | |
| "learning_rate": 1.564625850340136e-06, | |
| "loss": 0.4578, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.02374494652269587, | |
| "grad_norm": 0.7248512159636582, | |
| "learning_rate": 1.5782312925170067e-06, | |
| "loss": 0.48, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.023949644337546698, | |
| "grad_norm": 0.7545806311647761, | |
| "learning_rate": 1.5918367346938775e-06, | |
| "loss": 0.4884, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.024154342152397524, | |
| "grad_norm": 0.7683287783699582, | |
| "learning_rate": 1.6054421768707482e-06, | |
| "loss": 0.4834, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.02435903996724835, | |
| "grad_norm": 0.775426549385026, | |
| "learning_rate": 1.619047619047619e-06, | |
| "loss": 0.4822, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.024563737782099176, | |
| "grad_norm": 0.7149469826873975, | |
| "learning_rate": 1.6326530612244897e-06, | |
| "loss": 0.4721, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.02476843559695, | |
| "grad_norm": 0.7985393152422335, | |
| "learning_rate": 1.6462585034013606e-06, | |
| "loss": 0.5166, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.024973133411800828, | |
| "grad_norm": 0.7885248764092557, | |
| "learning_rate": 1.6598639455782312e-06, | |
| "loss": 0.5091, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.025177831226651657, | |
| "grad_norm": 0.7531097232781883, | |
| "learning_rate": 1.673469387755102e-06, | |
| "loss": 0.5133, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.025382529041502483, | |
| "grad_norm": 0.7097213698861701, | |
| "learning_rate": 1.6870748299319727e-06, | |
| "loss": 0.5001, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.02558722685635331, | |
| "grad_norm": 0.6936318152279768, | |
| "learning_rate": 1.7006802721088434e-06, | |
| "loss": 0.4611, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.025791924671204135, | |
| "grad_norm": 0.7442480820206602, | |
| "learning_rate": 1.714285714285714e-06, | |
| "loss": 0.5107, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.02599662248605496, | |
| "grad_norm": 0.7310368101162509, | |
| "learning_rate": 1.727891156462585e-06, | |
| "loss": 0.4568, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.026201320300905787, | |
| "grad_norm": 0.7723563494615043, | |
| "learning_rate": 1.7414965986394556e-06, | |
| "loss": 0.4976, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.026406018115756616, | |
| "grad_norm": 0.7688284872373655, | |
| "learning_rate": 1.7551020408163264e-06, | |
| "loss": 0.4876, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.026610715930607442, | |
| "grad_norm": 0.7663908612309938, | |
| "learning_rate": 1.768707482993197e-06, | |
| "loss": 0.5089, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.026815413745458268, | |
| "grad_norm": 0.6966352320510637, | |
| "learning_rate": 1.782312925170068e-06, | |
| "loss": 0.4537, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.027020111560309094, | |
| "grad_norm": 0.6933747179682217, | |
| "learning_rate": 1.7959183673469386e-06, | |
| "loss": 0.4431, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.02722480937515992, | |
| "grad_norm": 0.7620187715357651, | |
| "learning_rate": 1.8095238095238095e-06, | |
| "loss": 0.5002, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.027429507190010746, | |
| "grad_norm": 0.694229773433825, | |
| "learning_rate": 1.8231292517006801e-06, | |
| "loss": 0.4602, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.02763420500486157, | |
| "grad_norm": 0.7152627743695282, | |
| "learning_rate": 1.836734693877551e-06, | |
| "loss": 0.46, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.0278389028197124, | |
| "grad_norm": 0.7175467946942147, | |
| "learning_rate": 1.8503401360544217e-06, | |
| "loss": 0.4687, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.028043600634563227, | |
| "grad_norm": 0.7852808070086453, | |
| "learning_rate": 1.8639455782312925e-06, | |
| "loss": 0.5045, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.028248298449414053, | |
| "grad_norm": 0.7212069697520485, | |
| "learning_rate": 1.8775510204081632e-06, | |
| "loss": 0.458, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.02845299626426488, | |
| "grad_norm": 0.6901869666091209, | |
| "learning_rate": 1.891156462585034e-06, | |
| "loss": 0.4873, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.028657694079115704, | |
| "grad_norm": 0.7038286804084832, | |
| "learning_rate": 1.9047619047619045e-06, | |
| "loss": 0.4575, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.02886239189396653, | |
| "grad_norm": 0.7736514303776025, | |
| "learning_rate": 1.918367346938775e-06, | |
| "loss": 0.4989, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.02906708970881736, | |
| "grad_norm": 0.7546506061753928, | |
| "learning_rate": 1.931972789115646e-06, | |
| "loss": 0.6157, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.029271787523668186, | |
| "grad_norm": 0.7502287441885653, | |
| "learning_rate": 1.945578231292517e-06, | |
| "loss": 0.4744, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.02947648533851901, | |
| "grad_norm": 0.7544747394474504, | |
| "learning_rate": 1.9591836734693877e-06, | |
| "loss": 0.4652, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.029681183153369838, | |
| "grad_norm": 0.7400203790224271, | |
| "learning_rate": 1.972789115646258e-06, | |
| "loss": 0.4749, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.029885880968220663, | |
| "grad_norm": 0.7756604473816919, | |
| "learning_rate": 1.986394557823129e-06, | |
| "loss": 0.4879, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.03009057878307149, | |
| "grad_norm": 0.7364241267157726, | |
| "learning_rate": 2e-06, | |
| "loss": 0.4641, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.03029527659792232, | |
| "grad_norm": 0.7509999341558731, | |
| "learning_rate": 1.9999997801737146e-06, | |
| "loss": 0.4716, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.030499974412773145, | |
| "grad_norm": 0.7817167258395246, | |
| "learning_rate": 1.9999991206949555e-06, | |
| "loss": 0.478, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.03070467222762397, | |
| "grad_norm": 0.6975681554994494, | |
| "learning_rate": 1.9999980215640124e-06, | |
| "loss": 0.4698, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.030909370042474796, | |
| "grad_norm": 0.6671508819481775, | |
| "learning_rate": 1.9999964827813685e-06, | |
| "loss": 0.4502, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.031114067857325622, | |
| "grad_norm": 0.7588040820967348, | |
| "learning_rate": 1.9999945043477006e-06, | |
| "loss": 0.4932, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.03131876567217645, | |
| "grad_norm": 0.7918237654214221, | |
| "learning_rate": 1.9999920862638785e-06, | |
| "loss": 0.4676, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.031523463487027274, | |
| "grad_norm": 0.7313481708497578, | |
| "learning_rate": 1.999989228530965e-06, | |
| "loss": 0.458, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.031728161301878104, | |
| "grad_norm": 0.7363675329922608, | |
| "learning_rate": 1.9999859311502164e-06, | |
| "loss": 0.4794, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.031932859116728926, | |
| "grad_norm": 0.7511639384926047, | |
| "learning_rate": 1.999982194123083e-06, | |
| "loss": 0.4811, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.032137556931579755, | |
| "grad_norm": 0.6977601768137399, | |
| "learning_rate": 1.9999780174512074e-06, | |
| "loss": 0.5046, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.032342254746430585, | |
| "grad_norm": 0.6871160044462953, | |
| "learning_rate": 1.999973401136426e-06, | |
| "loss": 0.4473, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.03254695256128141, | |
| "grad_norm": 0.735276225575691, | |
| "learning_rate": 1.999968345180768e-06, | |
| "loss": 0.4769, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.03275165037613224, | |
| "grad_norm": 0.6707772270842888, | |
| "learning_rate": 1.999962849586457e-06, | |
| "loss": 0.4395, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.03295634819098306, | |
| "grad_norm": 0.7714337406838349, | |
| "learning_rate": 1.9999569143559085e-06, | |
| "loss": 0.4658, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.03316104600583389, | |
| "grad_norm": 0.7896606525524605, | |
| "learning_rate": 1.999950539491732e-06, | |
| "loss": 0.4645, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.03336574382068471, | |
| "grad_norm": 0.7359544675011239, | |
| "learning_rate": 1.999943724996731e-06, | |
| "loss": 0.4671, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.03357044163553554, | |
| "grad_norm": 0.6942155729771998, | |
| "learning_rate": 1.9999364708739005e-06, | |
| "loss": 0.4567, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.03377513945038637, | |
| "grad_norm": 0.7156881239994389, | |
| "learning_rate": 1.9999287771264305e-06, | |
| "loss": 0.4871, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.03397983726523719, | |
| "grad_norm": 0.7027330885247588, | |
| "learning_rate": 1.999920643757703e-06, | |
| "loss": 0.4371, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.03418453508008802, | |
| "grad_norm": 0.7022489082551948, | |
| "learning_rate": 1.9999120707712943e-06, | |
| "loss": 0.45, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.034389232894938844, | |
| "grad_norm": 0.7600832151805308, | |
| "learning_rate": 1.9999030581709736e-06, | |
| "loss": 0.4812, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.03459393070978967, | |
| "grad_norm": 0.7757365362216246, | |
| "learning_rate": 1.9998936059607028e-06, | |
| "loss": 0.4951, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.0347986285246405, | |
| "grad_norm": 0.7180624522326351, | |
| "learning_rate": 1.9998837141446378e-06, | |
| "loss": 0.4733, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.035003326339491325, | |
| "grad_norm": 0.8012677566963108, | |
| "learning_rate": 1.9998733827271277e-06, | |
| "loss": 0.4854, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.035208024154342155, | |
| "grad_norm": 0.715314370525801, | |
| "learning_rate": 1.999862611712715e-06, | |
| "loss": 0.4777, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.03541272196919298, | |
| "grad_norm": 0.7191382757417352, | |
| "learning_rate": 1.9998514011061344e-06, | |
| "loss": 0.4637, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.035617419784043806, | |
| "grad_norm": 0.7116753984408628, | |
| "learning_rate": 1.9998397509123154e-06, | |
| "loss": 0.4536, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.03582211759889463, | |
| "grad_norm": 0.7598078550909712, | |
| "learning_rate": 1.9998276611363797e-06, | |
| "loss": 0.4908, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.03602681541374546, | |
| "grad_norm": 0.6563039003765047, | |
| "learning_rate": 1.999815131783643e-06, | |
| "loss": 0.449, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.03623151322859629, | |
| "grad_norm": 0.719561949641505, | |
| "learning_rate": 1.999802162859613e-06, | |
| "loss": 0.4741, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.03643621104344711, | |
| "grad_norm": 0.7109902438469043, | |
| "learning_rate": 1.999788754369993e-06, | |
| "loss": 0.4701, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.03664090885829794, | |
| "grad_norm": 0.7065392449298251, | |
| "learning_rate": 1.9997749063206762e-06, | |
| "loss": 0.4714, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.03684560667314876, | |
| "grad_norm": 0.7696360740535267, | |
| "learning_rate": 1.9997606187177524e-06, | |
| "loss": 0.4875, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.03705030448799959, | |
| "grad_norm": 0.7305783700088637, | |
| "learning_rate": 1.999745891567502e-06, | |
| "loss": 0.4606, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.037255002302850414, | |
| "grad_norm": 0.7270975727384246, | |
| "learning_rate": 1.9997307248764014e-06, | |
| "loss": 0.4198, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.03745970011770124, | |
| "grad_norm": 0.7569285986642791, | |
| "learning_rate": 1.9997151186511173e-06, | |
| "loss": 0.4354, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.03766439793255207, | |
| "grad_norm": 0.7219339206651326, | |
| "learning_rate": 1.9996990728985115e-06, | |
| "loss": 0.4378, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.037869095747402895, | |
| "grad_norm": 0.7690405011750759, | |
| "learning_rate": 1.9996825876256386e-06, | |
| "loss": 0.4791, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.038073793562253724, | |
| "grad_norm": 0.7552362068529521, | |
| "learning_rate": 1.9996656628397466e-06, | |
| "loss": 0.4672, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.03827849137710455, | |
| "grad_norm": 0.7341580278198813, | |
| "learning_rate": 1.999648298548276e-06, | |
| "loss": 0.4677, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.038483189191955376, | |
| "grad_norm": 0.7067121453226938, | |
| "learning_rate": 1.9996304947588612e-06, | |
| "loss": 0.4727, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.038687887006806206, | |
| "grad_norm": 0.7237165727925357, | |
| "learning_rate": 1.99961225147933e-06, | |
| "loss": 0.4446, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.03889258482165703, | |
| "grad_norm": 0.7069486805133093, | |
| "learning_rate": 1.999593568717703e-06, | |
| "loss": 0.4599, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.03909728263650786, | |
| "grad_norm": 0.897481774030034, | |
| "learning_rate": 1.9995744464821936e-06, | |
| "loss": 0.5129, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.03930198045135868, | |
| "grad_norm": 0.7488636278687589, | |
| "learning_rate": 1.9995548847812097e-06, | |
| "loss": 0.5184, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.03950667826620951, | |
| "grad_norm": 0.7194583225483666, | |
| "learning_rate": 1.9995348836233515e-06, | |
| "loss": 0.4915, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.03971137608106033, | |
| "grad_norm": 0.7384378767131218, | |
| "learning_rate": 1.999514443017412e-06, | |
| "loss": 0.4487, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.03991607389591116, | |
| "grad_norm": 0.7577332351147034, | |
| "learning_rate": 1.9994935629723784e-06, | |
| "loss": 0.4842, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.04012077171076199, | |
| "grad_norm": 0.7207002083905842, | |
| "learning_rate": 1.999472243497431e-06, | |
| "loss": 0.4698, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.04032546952561281, | |
| "grad_norm": 0.8010492120535461, | |
| "learning_rate": 1.9994504846019423e-06, | |
| "loss": 0.4561, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.04053016734046364, | |
| "grad_norm": 0.7453701461541147, | |
| "learning_rate": 1.9994282862954787e-06, | |
| "loss": 0.4806, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.040734865155314465, | |
| "grad_norm": 0.7255193966716207, | |
| "learning_rate": 1.9994056485878002e-06, | |
| "loss": 0.4511, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.040939562970165294, | |
| "grad_norm": 0.7957588909816856, | |
| "learning_rate": 1.9993825714888594e-06, | |
| "loss": 0.4775, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.04114426078501612, | |
| "grad_norm": 0.7304580504624026, | |
| "learning_rate": 1.999359055008802e-06, | |
| "loss": 0.4476, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.041348958599866946, | |
| "grad_norm": 0.8052138479705295, | |
| "learning_rate": 1.999335099157967e-06, | |
| "loss": 0.4621, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.041553656414717775, | |
| "grad_norm": 0.7344879094324241, | |
| "learning_rate": 1.999310703946887e-06, | |
| "loss": 0.448, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.0417583542295686, | |
| "grad_norm": 0.804269507197302, | |
| "learning_rate": 1.999285869386287e-06, | |
| "loss": 0.471, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.04196305204441943, | |
| "grad_norm": 0.7284627322104599, | |
| "learning_rate": 1.9992605954870867e-06, | |
| "loss": 0.4418, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.04216774985927025, | |
| "grad_norm": 0.7243013667651625, | |
| "learning_rate": 1.999234882260396e-06, | |
| "loss": 0.4669, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.04237244767412108, | |
| "grad_norm": 0.677583318692503, | |
| "learning_rate": 1.9992087297175213e-06, | |
| "loss": 0.4447, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.04257714548897191, | |
| "grad_norm": 0.7334595699121094, | |
| "learning_rate": 1.9991821378699598e-06, | |
| "loss": 0.4719, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.04278184330382273, | |
| "grad_norm": 0.7351912069847943, | |
| "learning_rate": 1.999155106729403e-06, | |
| "loss": 0.4758, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.04298654111867356, | |
| "grad_norm": 0.7262994043092325, | |
| "learning_rate": 1.9991276363077355e-06, | |
| "loss": 0.4636, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.04319123893352438, | |
| "grad_norm": 0.7170624975773432, | |
| "learning_rate": 1.999099726617034e-06, | |
| "loss": 0.4432, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.04339593674837521, | |
| "grad_norm": 0.7756861925710989, | |
| "learning_rate": 1.9990713776695697e-06, | |
| "loss": 0.4277, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.043600634563226034, | |
| "grad_norm": 0.7766290291464314, | |
| "learning_rate": 1.999042589477806e-06, | |
| "loss": 0.4521, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.043805332378076864, | |
| "grad_norm": 0.7675003445260637, | |
| "learning_rate": 1.9990133620543992e-06, | |
| "loss": 0.4728, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.04401003019292769, | |
| "grad_norm": 0.7502537566865808, | |
| "learning_rate": 1.9989836954122006e-06, | |
| "loss": 0.4919, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.044214728007778516, | |
| "grad_norm": 0.7256221777073304, | |
| "learning_rate": 1.998953589564252e-06, | |
| "loss": 0.4427, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.044419425822629345, | |
| "grad_norm": 0.7209802907423725, | |
| "learning_rate": 1.9989230445237905e-06, | |
| "loss": 0.4482, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.04462412363748017, | |
| "grad_norm": 0.761848969478383, | |
| "learning_rate": 1.9988920603042437e-06, | |
| "loss": 0.4623, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.044828821452331, | |
| "grad_norm": 0.7511377700619639, | |
| "learning_rate": 1.9988606369192357e-06, | |
| "loss": 0.4695, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.045033519267181826, | |
| "grad_norm": 0.684910692983434, | |
| "learning_rate": 1.998828774382581e-06, | |
| "loss": 0.4546, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.04523821708203265, | |
| "grad_norm": 0.8229255435418116, | |
| "learning_rate": 1.998796472708288e-06, | |
| "loss": 0.4736, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.04544291489688348, | |
| "grad_norm": 0.7208349457907924, | |
| "learning_rate": 1.998763731910558e-06, | |
| "loss": 0.4464, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.0456476127117343, | |
| "grad_norm": 0.8196660504458043, | |
| "learning_rate": 1.998730552003786e-06, | |
| "loss": 0.5129, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.04585231052658513, | |
| "grad_norm": 0.7410783352083353, | |
| "learning_rate": 1.99869693300256e-06, | |
| "loss": 0.4716, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.04605700834143595, | |
| "grad_norm": 0.6829979843011463, | |
| "learning_rate": 1.9986628749216598e-06, | |
| "loss": 0.4543, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.04626170615628678, | |
| "grad_norm": 0.6883777439066587, | |
| "learning_rate": 1.9986283777760598e-06, | |
| "loss": 0.4441, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.04646640397113761, | |
| "grad_norm": 0.7411624814334564, | |
| "learning_rate": 1.9985934415809266e-06, | |
| "loss": 0.4902, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.046671101785988434, | |
| "grad_norm": 0.7207442928316582, | |
| "learning_rate": 1.99855806635162e-06, | |
| "loss": 0.4502, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.04687579960083926, | |
| "grad_norm": 0.7535800049970842, | |
| "learning_rate": 1.9985222521036923e-06, | |
| "loss": 0.4588, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.047080497415690085, | |
| "grad_norm": 0.755131046087142, | |
| "learning_rate": 1.9984859988528896e-06, | |
| "loss": 0.478, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.047285195230540915, | |
| "grad_norm": 0.7363267944788255, | |
| "learning_rate": 1.9984493066151515e-06, | |
| "loss": 0.4722, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.04748989304539174, | |
| "grad_norm": 0.7175156333070826, | |
| "learning_rate": 1.9984121754066084e-06, | |
| "loss": 0.4284, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.04769459086024257, | |
| "grad_norm": 0.7021868758602576, | |
| "learning_rate": 1.9983746052435867e-06, | |
| "loss": 0.4549, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.047899288675093396, | |
| "grad_norm": 0.7661651319293146, | |
| "learning_rate": 1.998336596142603e-06, | |
| "loss": 0.4626, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.04810398648994422, | |
| "grad_norm": 0.7200117942966474, | |
| "learning_rate": 1.9982981481203685e-06, | |
| "loss": 0.4602, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.04830868430479505, | |
| "grad_norm": 0.7402636115111145, | |
| "learning_rate": 1.9982592611937875e-06, | |
| "loss": 0.462, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.04851338211964587, | |
| "grad_norm": 0.7289521015317652, | |
| "learning_rate": 1.998219935379956e-06, | |
| "loss": 0.4247, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.0487180799344967, | |
| "grad_norm": 0.8254564955090967, | |
| "learning_rate": 1.9981801706961637e-06, | |
| "loss": 0.445, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.04892277774934753, | |
| "grad_norm": 0.711987818470011, | |
| "learning_rate": 1.9981399671598938e-06, | |
| "loss": 0.4373, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.04912747556419835, | |
| "grad_norm": 0.7578069994316992, | |
| "learning_rate": 1.9980993247888215e-06, | |
| "loss": 0.4397, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.04933217337904918, | |
| "grad_norm": 0.7111156315088227, | |
| "learning_rate": 1.9980582436008155e-06, | |
| "loss": 0.444, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.0495368711939, | |
| "grad_norm": 0.6825451116028749, | |
| "learning_rate": 1.998016723613937e-06, | |
| "loss": 0.4216, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.04974156900875083, | |
| "grad_norm": 0.717357311135891, | |
| "learning_rate": 1.9979747648464406e-06, | |
| "loss": 0.4393, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.049946266823601655, | |
| "grad_norm": 0.7431878666336732, | |
| "learning_rate": 1.9979323673167735e-06, | |
| "loss": 0.4588, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.050150964638452485, | |
| "grad_norm": 0.7393037207206594, | |
| "learning_rate": 1.997889531043576e-06, | |
| "loss": 0.4547, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.050355662453303314, | |
| "grad_norm": 0.751624588819876, | |
| "learning_rate": 1.997846256045681e-06, | |
| "loss": 0.4423, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.050560360268154136, | |
| "grad_norm": 0.7291703435937729, | |
| "learning_rate": 1.9978025423421143e-06, | |
| "loss": 0.4577, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.050765058083004966, | |
| "grad_norm": 0.7730058435275767, | |
| "learning_rate": 1.9977583899520954e-06, | |
| "loss": 0.4725, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.05096975589785579, | |
| "grad_norm": 0.7388265631923211, | |
| "learning_rate": 1.9977137988950354e-06, | |
| "loss": 0.4915, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.05117445371270662, | |
| "grad_norm": 0.7253056655402751, | |
| "learning_rate": 1.9976687691905393e-06, | |
| "loss": 0.4463, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.05137915152755744, | |
| "grad_norm": 0.7279513578282504, | |
| "learning_rate": 1.997623300858404e-06, | |
| "loss": 0.4692, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.05158384934240827, | |
| "grad_norm": 0.7758619445867678, | |
| "learning_rate": 1.99757739391862e-06, | |
| "loss": 0.4359, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.0517885471572591, | |
| "grad_norm": 0.6934936677414176, | |
| "learning_rate": 1.9975310483913706e-06, | |
| "loss": 0.4342, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.05199324497210992, | |
| "grad_norm": 0.785492944076531, | |
| "learning_rate": 1.9974842642970316e-06, | |
| "loss": 0.4762, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.05219794278696075, | |
| "grad_norm": 0.693346587621168, | |
| "learning_rate": 1.9974370416561716e-06, | |
| "loss": 0.4077, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.05240264060181157, | |
| "grad_norm": 0.7036807587452536, | |
| "learning_rate": 1.9973893804895526e-06, | |
| "loss": 0.4559, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.0526073384166624, | |
| "grad_norm": 0.6487321250079171, | |
| "learning_rate": 1.997341280818128e-06, | |
| "loss": 0.4445, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.05281203623151323, | |
| "grad_norm": 0.7581793864091325, | |
| "learning_rate": 1.9972927426630464e-06, | |
| "loss": 0.4189, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.053016734046364054, | |
| "grad_norm": 0.7654003299344445, | |
| "learning_rate": 1.9972437660456465e-06, | |
| "loss": 0.4772, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.053221431861214884, | |
| "grad_norm": 0.7482689765950153, | |
| "learning_rate": 1.9971943509874614e-06, | |
| "loss": 0.4577, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.053426129676065706, | |
| "grad_norm": 0.7328225919609754, | |
| "learning_rate": 1.997144497510217e-06, | |
| "loss": 0.4301, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.053630827490916536, | |
| "grad_norm": 0.734577771662883, | |
| "learning_rate": 1.9970942056358307e-06, | |
| "loss": 0.4721, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.05383552530576736, | |
| "grad_norm": 0.7974471697046129, | |
| "learning_rate": 1.997043475386414e-06, | |
| "loss": 0.4759, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.05404022312061819, | |
| "grad_norm": 0.7204102805022299, | |
| "learning_rate": 1.99699230678427e-06, | |
| "loss": 0.4159, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.05424492093546902, | |
| "grad_norm": 0.8012057502786673, | |
| "learning_rate": 1.996940699851896e-06, | |
| "loss": 0.4784, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.05444961875031984, | |
| "grad_norm": 0.7456181199531785, | |
| "learning_rate": 1.9968886546119805e-06, | |
| "loss": 0.4716, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.05465431656517067, | |
| "grad_norm": 0.7582815001255205, | |
| "learning_rate": 1.996836171087405e-06, | |
| "loss": 0.4561, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.05485901438002149, | |
| "grad_norm": 0.7334168776176787, | |
| "learning_rate": 1.996783249301245e-06, | |
| "loss": 0.4344, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.05506371219487232, | |
| "grad_norm": 0.6872926048341936, | |
| "learning_rate": 1.9967298892767674e-06, | |
| "loss": 0.4293, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.05526841000972314, | |
| "grad_norm": 0.7532210670290133, | |
| "learning_rate": 1.9966760910374313e-06, | |
| "loss": 0.4644, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.05547310782457397, | |
| "grad_norm": 0.7306067456669837, | |
| "learning_rate": 1.99662185460689e-06, | |
| "loss": 0.4618, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.0556778056394248, | |
| "grad_norm": 0.7456313505165526, | |
| "learning_rate": 1.9965671800089887e-06, | |
| "loss": 0.4389, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.055882503454275624, | |
| "grad_norm": 0.6963357390921938, | |
| "learning_rate": 1.9965120672677646e-06, | |
| "loss": 0.4596, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.056087201269126453, | |
| "grad_norm": 0.6963762346496246, | |
| "learning_rate": 1.9964565164074488e-06, | |
| "loss": 0.4452, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.056291899083977276, | |
| "grad_norm": 0.7538437049515703, | |
| "learning_rate": 1.996400527452464e-06, | |
| "loss": 0.442, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.056496596898828105, | |
| "grad_norm": 0.7129437285411915, | |
| "learning_rate": 1.9963441004274265e-06, | |
| "loss": 0.4575, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.056701294713678935, | |
| "grad_norm": 0.7500374688918953, | |
| "learning_rate": 1.9962872353571436e-06, | |
| "loss": 0.4739, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.05690599252852976, | |
| "grad_norm": 0.7695931372733311, | |
| "learning_rate": 1.996229932266617e-06, | |
| "loss": 0.4523, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.05711069034338059, | |
| "grad_norm": 0.8047384042314083, | |
| "learning_rate": 1.99617219118104e-06, | |
| "loss": 0.4541, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.05731538815823141, | |
| "grad_norm": 0.7337412326319969, | |
| "learning_rate": 1.9961140121257978e-06, | |
| "loss": 0.5429, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.05752008597308224, | |
| "grad_norm": 0.7647644479794993, | |
| "learning_rate": 1.99605539512647e-06, | |
| "loss": 0.4479, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.05772478378793306, | |
| "grad_norm": 0.7676226046817726, | |
| "learning_rate": 1.9959963402088274e-06, | |
| "loss": 0.4641, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.05792948160278389, | |
| "grad_norm": 0.7235143819897621, | |
| "learning_rate": 1.9959368473988333e-06, | |
| "loss": 0.4545, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.05813417941763472, | |
| "grad_norm": 0.7247778707795571, | |
| "learning_rate": 1.9958769167226444e-06, | |
| "loss": 0.4527, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.05833887723248554, | |
| "grad_norm": 0.7539665158584474, | |
| "learning_rate": 1.995816548206609e-06, | |
| "loss": 0.4458, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.05854357504733637, | |
| "grad_norm": 0.7532087725317325, | |
| "learning_rate": 1.995755741877269e-06, | |
| "loss": 0.4461, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.058748272862187194, | |
| "grad_norm": 0.7797916526020292, | |
| "learning_rate": 1.995694497761357e-06, | |
| "loss": 0.4665, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.05895297067703802, | |
| "grad_norm": 0.7517569918058703, | |
| "learning_rate": 1.9956328158857992e-06, | |
| "loss": 0.4728, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.059157668491888846, | |
| "grad_norm": 0.7570745421432602, | |
| "learning_rate": 1.995570696277715e-06, | |
| "loss": 0.4563, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.059362366306739675, | |
| "grad_norm": 0.740586734451177, | |
| "learning_rate": 1.995508138964415e-06, | |
| "loss": 0.4709, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.059567064121590504, | |
| "grad_norm": 0.7681915361666245, | |
| "learning_rate": 1.995445143973403e-06, | |
| "loss": 0.4503, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.05977176193644133, | |
| "grad_norm": 0.7455448750612553, | |
| "learning_rate": 1.9953817113323743e-06, | |
| "loss": 0.4529, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.059976459751292156, | |
| "grad_norm": 0.7745938996976398, | |
| "learning_rate": 1.9953178410692174e-06, | |
| "loss": 0.4256, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.06018115756614298, | |
| "grad_norm": 0.7431821602019313, | |
| "learning_rate": 1.9952535332120137e-06, | |
| "loss": 0.4453, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.06038585538099381, | |
| "grad_norm": 0.6903595364669262, | |
| "learning_rate": 1.9951887877890354e-06, | |
| "loss": 0.4339, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.06059055319584464, | |
| "grad_norm": 0.773597292773469, | |
| "learning_rate": 1.9951236048287483e-06, | |
| "loss": 0.4817, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.06079525101069546, | |
| "grad_norm": 0.7263693618591641, | |
| "learning_rate": 1.9950579843598105e-06, | |
| "loss": 0.4572, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.06099994882554629, | |
| "grad_norm": 0.7297961413486055, | |
| "learning_rate": 1.994991926411072e-06, | |
| "loss": 0.4359, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.06120464664039711, | |
| "grad_norm": 0.7855052727291876, | |
| "learning_rate": 1.9949254310115753e-06, | |
| "loss": 0.4568, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.06140934445524794, | |
| "grad_norm": 0.7300504368627593, | |
| "learning_rate": 1.994858498190556e-06, | |
| "loss": 0.4501, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.061614042270098764, | |
| "grad_norm": 0.7096161782700376, | |
| "learning_rate": 1.99479112797744e-06, | |
| "loss": 0.4663, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.06181874008494959, | |
| "grad_norm": 0.7018494412530502, | |
| "learning_rate": 1.9947233204018477e-06, | |
| "loss": 0.4401, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.06202343789980042, | |
| "grad_norm": 0.7907086687325865, | |
| "learning_rate": 1.9946550754935906e-06, | |
| "loss": 0.4394, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.062228135714651245, | |
| "grad_norm": 0.7637305795975494, | |
| "learning_rate": 1.9945863932826727e-06, | |
| "loss": 0.4368, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.062432833529502074, | |
| "grad_norm": 0.7745057196668337, | |
| "learning_rate": 1.9945172737992904e-06, | |
| "loss": 0.4926, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.0626375313443529, | |
| "grad_norm": 0.7367608831304401, | |
| "learning_rate": 1.994447717073832e-06, | |
| "loss": 0.4688, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.06284222915920372, | |
| "grad_norm": 0.7695856962055981, | |
| "learning_rate": 1.9943777231368794e-06, | |
| "loss": 0.4484, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.06304692697405455, | |
| "grad_norm": 0.7187776209986876, | |
| "learning_rate": 1.994307292019204e-06, | |
| "loss": 0.4444, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.06325162478890538, | |
| "grad_norm": 0.7347949665208309, | |
| "learning_rate": 1.994236423751772e-06, | |
| "loss": 0.4485, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.06345632260375621, | |
| "grad_norm": 0.7105615411584904, | |
| "learning_rate": 1.99416511836574e-06, | |
| "loss": 0.4115, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.06366102041860704, | |
| "grad_norm": 0.7092446318381356, | |
| "learning_rate": 1.994093375892459e-06, | |
| "loss": 0.4398, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.06386571823345785, | |
| "grad_norm": 0.770807237418002, | |
| "learning_rate": 1.9940211963634696e-06, | |
| "loss": 0.4413, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.06407041604830868, | |
| "grad_norm": 0.7396584102001305, | |
| "learning_rate": 1.9939485798105057e-06, | |
| "loss": 0.4851, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.06427511386315951, | |
| "grad_norm": 0.7444518429400907, | |
| "learning_rate": 1.9938755262654945e-06, | |
| "loss": 0.4337, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.06447981167801034, | |
| "grad_norm": 0.722720227886292, | |
| "learning_rate": 1.9938020357605527e-06, | |
| "loss": 0.4965, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.06468450949286117, | |
| "grad_norm": 0.7666649202795676, | |
| "learning_rate": 1.993728108327992e-06, | |
| "loss": 0.4532, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.06488920730771199, | |
| "grad_norm": 0.7549308832804502, | |
| "learning_rate": 1.9936537440003134e-06, | |
| "loss": 0.4582, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.06509390512256281, | |
| "grad_norm": 0.6872546619483418, | |
| "learning_rate": 1.993578942810212e-06, | |
| "loss": 0.4695, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.06529860293741364, | |
| "grad_norm": 0.7550879312423509, | |
| "learning_rate": 1.9935037047905748e-06, | |
| "loss": 0.4833, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.06550330075226447, | |
| "grad_norm": 0.7064631296777315, | |
| "learning_rate": 1.99342802997448e-06, | |
| "loss": 0.4724, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.0657079985671153, | |
| "grad_norm": 0.7041398494235039, | |
| "learning_rate": 1.9933519183951977e-06, | |
| "loss": 0.4441, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.06591269638196612, | |
| "grad_norm": 0.7624056534749137, | |
| "learning_rate": 1.9932753700861914e-06, | |
| "loss": 0.465, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.06611739419681695, | |
| "grad_norm": 0.7553645719854849, | |
| "learning_rate": 1.9931983850811155e-06, | |
| "loss": 0.4241, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.06632209201166778, | |
| "grad_norm": 0.7617302308064162, | |
| "learning_rate": 1.9931209634138158e-06, | |
| "loss": 0.4475, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.0665267898265186, | |
| "grad_norm": 0.7484410349937703, | |
| "learning_rate": 1.9930431051183324e-06, | |
| "loss": 0.432, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.06673148764136942, | |
| "grad_norm": 0.7283423040518047, | |
| "learning_rate": 1.9929648102288953e-06, | |
| "loss": 0.4388, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.06693618545622025, | |
| "grad_norm": 0.7852619772614368, | |
| "learning_rate": 1.9928860787799265e-06, | |
| "loss": 0.468, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.06714088327107108, | |
| "grad_norm": 0.7769054079891817, | |
| "learning_rate": 1.992806910806041e-06, | |
| "loss": 0.4579, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.06734558108592191, | |
| "grad_norm": 0.7619008198687942, | |
| "learning_rate": 1.992727306342045e-06, | |
| "loss": 0.4789, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.06755027890077274, | |
| "grad_norm": 0.7607061858221339, | |
| "learning_rate": 1.9926472654229376e-06, | |
| "loss": 0.4241, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.06775497671562355, | |
| "grad_norm": 0.7236097758888326, | |
| "learning_rate": 1.992566788083908e-06, | |
| "loss": 0.4181, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.06795967453047438, | |
| "grad_norm": 0.8114241422692142, | |
| "learning_rate": 1.992485874360338e-06, | |
| "loss": 0.4383, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.06816437234532521, | |
| "grad_norm": 0.7429107582072085, | |
| "learning_rate": 1.992404524287803e-06, | |
| "loss": 0.4418, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.06836907016017604, | |
| "grad_norm": 0.7664592844718724, | |
| "learning_rate": 1.9923227379020674e-06, | |
| "loss": 0.4424, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.06857376797502687, | |
| "grad_norm": 0.7525421136101852, | |
| "learning_rate": 1.9922405152390893e-06, | |
| "loss": 0.4601, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.06877846578987769, | |
| "grad_norm": 0.7924977816871237, | |
| "learning_rate": 1.9921578563350182e-06, | |
| "loss": 0.4481, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.06898316360472852, | |
| "grad_norm": 0.7205455607720451, | |
| "learning_rate": 1.9920747612261953e-06, | |
| "loss": 0.472, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.06918786141957935, | |
| "grad_norm": 0.8099842191878124, | |
| "learning_rate": 1.9919912299491534e-06, | |
| "loss": 0.4522, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.06939255923443018, | |
| "grad_norm": 0.7601204829884295, | |
| "learning_rate": 1.991907262540617e-06, | |
| "loss": 0.4361, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.069597257049281, | |
| "grad_norm": 0.7498165219729875, | |
| "learning_rate": 1.9918228590375034e-06, | |
| "loss": 0.4363, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.06980195486413182, | |
| "grad_norm": 0.7130352034738562, | |
| "learning_rate": 1.9917380194769197e-06, | |
| "loss": 0.4355, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.07000665267898265, | |
| "grad_norm": 0.7653650161746682, | |
| "learning_rate": 1.991652743896167e-06, | |
| "loss": 0.4062, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.07021135049383348, | |
| "grad_norm": 0.8196206368298663, | |
| "learning_rate": 1.991567032332736e-06, | |
| "loss": 0.4614, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.07041604830868431, | |
| "grad_norm": 0.7812062890731587, | |
| "learning_rate": 1.991480884824311e-06, | |
| "loss": 0.4975, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.07062074612353512, | |
| "grad_norm": 0.7206276352235353, | |
| "learning_rate": 1.9913943014087655e-06, | |
| "loss": 0.4157, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.07082544393838595, | |
| "grad_norm": 0.78698897087266, | |
| "learning_rate": 1.9913072821241672e-06, | |
| "loss": 0.4581, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.07103014175323678, | |
| "grad_norm": 0.7764210083677198, | |
| "learning_rate": 1.991219827008775e-06, | |
| "loss": 0.4408, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.07123483956808761, | |
| "grad_norm": 0.7595433195005857, | |
| "learning_rate": 1.9911319361010367e-06, | |
| "loss": 0.4492, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.07143953738293844, | |
| "grad_norm": 0.722334537399672, | |
| "learning_rate": 1.991043609439596e-06, | |
| "loss": 0.4098, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.07164423519778926, | |
| "grad_norm": 0.7424362807959037, | |
| "learning_rate": 1.9909548470632842e-06, | |
| "loss": 0.4169, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.07184893301264009, | |
| "grad_norm": 0.7665542829825039, | |
| "learning_rate": 1.9908656490111267e-06, | |
| "loss": 0.4635, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.07205363082749092, | |
| "grad_norm": 0.7670493050777013, | |
| "learning_rate": 1.9907760153223396e-06, | |
| "loss": 0.4245, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.07225832864234175, | |
| "grad_norm": 0.7082306081348982, | |
| "learning_rate": 1.9906859460363304e-06, | |
| "loss": 0.3885, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.07246302645719258, | |
| "grad_norm": 0.7179046673862461, | |
| "learning_rate": 1.990595441192699e-06, | |
| "loss": 0.4047, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.07266772427204339, | |
| "grad_norm": 0.7785673633417279, | |
| "learning_rate": 1.990504500831235e-06, | |
| "loss": 0.484, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.07287242208689422, | |
| "grad_norm": 0.7657292926077239, | |
| "learning_rate": 1.9904131249919215e-06, | |
| "loss": 0.4444, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.07307711990174505, | |
| "grad_norm": 0.7569753573050043, | |
| "learning_rate": 1.9903213137149313e-06, | |
| "loss": 0.4701, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.07328181771659588, | |
| "grad_norm": 0.8072244283837906, | |
| "learning_rate": 1.99022906704063e-06, | |
| "loss": 0.4409, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.07348651553144671, | |
| "grad_norm": 0.7629374834771911, | |
| "learning_rate": 1.990136385009574e-06, | |
| "loss": 0.4927, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.07369121334629752, | |
| "grad_norm": 0.7385249798311992, | |
| "learning_rate": 1.990043267662511e-06, | |
| "loss": 0.4338, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.07389591116114835, | |
| "grad_norm": 0.775227359922002, | |
| "learning_rate": 1.989949715040381e-06, | |
| "loss": 0.4789, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.07410060897599918, | |
| "grad_norm": 0.7517526442766227, | |
| "learning_rate": 1.9898557271843133e-06, | |
| "loss": 0.4504, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.07430530679085001, | |
| "grad_norm": 0.7375927174293256, | |
| "learning_rate": 1.9897613041356314e-06, | |
| "loss": 0.4108, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.07451000460570083, | |
| "grad_norm": 0.72174871971371, | |
| "learning_rate": 1.9896664459358472e-06, | |
| "loss": 0.4475, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.07471470242055166, | |
| "grad_norm": 0.7329776456768429, | |
| "learning_rate": 1.9895711526266667e-06, | |
| "loss": 0.424, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.07491940023540249, | |
| "grad_norm": 0.7185696661034995, | |
| "learning_rate": 1.9894754242499852e-06, | |
| "loss": 0.4543, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.07512409805025332, | |
| "grad_norm": 0.7049184644044292, | |
| "learning_rate": 1.98937926084789e-06, | |
| "loss": 0.42, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.07532879586510415, | |
| "grad_norm": 0.7293724081932391, | |
| "learning_rate": 1.989282662462659e-06, | |
| "loss": 0.43, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.07553349367995496, | |
| "grad_norm": 0.7059632003848231, | |
| "learning_rate": 1.9891856291367626e-06, | |
| "loss": 0.4275, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.07573819149480579, | |
| "grad_norm": 0.7683789138077434, | |
| "learning_rate": 1.9890881609128618e-06, | |
| "loss": 0.4408, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.07594288930965662, | |
| "grad_norm": 0.7872951646579199, | |
| "learning_rate": 1.9889902578338087e-06, | |
| "loss": 0.4292, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.07614758712450745, | |
| "grad_norm": 0.7924075912782186, | |
| "learning_rate": 1.988891919942646e-06, | |
| "loss": 0.4408, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.07635228493935828, | |
| "grad_norm": 0.7607074410846072, | |
| "learning_rate": 1.9887931472826093e-06, | |
| "loss": 0.4485, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.0765569827542091, | |
| "grad_norm": 0.7592064137840062, | |
| "learning_rate": 1.9886939398971238e-06, | |
| "loss": 0.4404, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.07676168056905992, | |
| "grad_norm": 0.8183614522835663, | |
| "learning_rate": 1.9885942978298054e-06, | |
| "loss": 0.4677, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.07696637838391075, | |
| "grad_norm": 0.7400985748589023, | |
| "learning_rate": 1.9884942211244637e-06, | |
| "loss": 0.4867, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.07717107619876158, | |
| "grad_norm": 0.7657568826393069, | |
| "learning_rate": 1.988393709825096e-06, | |
| "loss": 0.4592, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.07737577401361241, | |
| "grad_norm": 0.7242932934680155, | |
| "learning_rate": 1.988292763975893e-06, | |
| "loss": 0.4037, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.07758047182846323, | |
| "grad_norm": 0.695616397865644, | |
| "learning_rate": 1.9881913836212365e-06, | |
| "loss": 0.4534, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.07778516964331406, | |
| "grad_norm": 0.7286084983103666, | |
| "learning_rate": 1.9880895688056977e-06, | |
| "loss": 0.4267, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.07798986745816489, | |
| "grad_norm": 0.7434704183939207, | |
| "learning_rate": 1.98798731957404e-06, | |
| "loss": 0.4308, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.07819456527301571, | |
| "grad_norm": 0.7803766550265465, | |
| "learning_rate": 1.9878846359712176e-06, | |
| "loss": 0.4455, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.07839926308786653, | |
| "grad_norm": 0.777748381274515, | |
| "learning_rate": 1.9877815180423757e-06, | |
| "loss": 0.4639, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.07860396090271736, | |
| "grad_norm": 0.724760659067026, | |
| "learning_rate": 1.9876779658328503e-06, | |
| "loss": 0.4666, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.07880865871756819, | |
| "grad_norm": 0.6620651646162441, | |
| "learning_rate": 1.9875739793881685e-06, | |
| "loss": 0.4195, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.07901335653241902, | |
| "grad_norm": 0.7524415209772658, | |
| "learning_rate": 1.9874695587540477e-06, | |
| "loss": 0.4431, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.07921805434726985, | |
| "grad_norm": 0.6797630439484897, | |
| "learning_rate": 1.9873647039763975e-06, | |
| "loss": 0.4453, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.07942275216212066, | |
| "grad_norm": 0.7685812672797145, | |
| "learning_rate": 1.987259415101317e-06, | |
| "loss": 0.4623, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.07962744997697149, | |
| "grad_norm": 0.7581895406953137, | |
| "learning_rate": 1.9871536921750965e-06, | |
| "loss": 0.4423, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.07983214779182232, | |
| "grad_norm": 0.7100776210964536, | |
| "learning_rate": 1.987047535244218e-06, | |
| "loss": 0.4171, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.08003684560667315, | |
| "grad_norm": 0.7299757652818287, | |
| "learning_rate": 1.9869409443553535e-06, | |
| "loss": 0.454, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.08024154342152398, | |
| "grad_norm": 0.7439290629641574, | |
| "learning_rate": 1.9868339195553657e-06, | |
| "loss": 0.4426, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.0804462412363748, | |
| "grad_norm": 0.75716235792549, | |
| "learning_rate": 1.9867264608913084e-06, | |
| "loss": 0.4479, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.08065093905122563, | |
| "grad_norm": 0.6968938261452492, | |
| "learning_rate": 1.9866185684104266e-06, | |
| "loss": 0.4335, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.08085563686607646, | |
| "grad_norm": 0.7170546940871543, | |
| "learning_rate": 1.9865102421601545e-06, | |
| "loss": 0.4286, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.08106033468092728, | |
| "grad_norm": 0.7478223773320385, | |
| "learning_rate": 1.986401482188119e-06, | |
| "loss": 0.4202, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.08126503249577811, | |
| "grad_norm": 0.7076328746009948, | |
| "learning_rate": 1.986292288542136e-06, | |
| "loss": 0.4558, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.08146973031062893, | |
| "grad_norm": 0.7591423650217918, | |
| "learning_rate": 1.986182661270213e-06, | |
| "loss": 0.4674, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.08167442812547976, | |
| "grad_norm": 0.7416281412961139, | |
| "learning_rate": 1.9860726004205485e-06, | |
| "loss": 0.4397, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.08187912594033059, | |
| "grad_norm": 0.7416913042041706, | |
| "learning_rate": 1.98596210604153e-06, | |
| "loss": 0.4405, | |
| "step": 400 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 4885, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 200, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 53820235284480.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |