| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.000483392030937, | |
| "global_step": 14484, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.9982733614199875e-05, | |
| "loss": 3.3505, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.994820084259963e-05, | |
| "loss": 1.7948, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.991366807099938e-05, | |
| "loss": 1.5872, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.9879135299399134e-05, | |
| "loss": 1.5076, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 4.984460252779889e-05, | |
| "loss": 1.4722, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.981006975619863e-05, | |
| "loss": 1.4505, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.9775536984598386e-05, | |
| "loss": 1.44, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.974100421299814e-05, | |
| "loss": 1.407, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 4.970647144139789e-05, | |
| "loss": 1.4086, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.9671938669797644e-05, | |
| "loss": 1.3679, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.963740589819739e-05, | |
| "loss": 1.3867, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.960287312659714e-05, | |
| "loss": 1.3467, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.9568340354996896e-05, | |
| "loss": 1.3652, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.953380758339665e-05, | |
| "loss": 1.3433, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.9499274811796395e-05, | |
| "loss": 1.3221, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.946474204019615e-05, | |
| "loss": 1.3264, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.94302092685959e-05, | |
| "loss": 1.3122, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.939567649699565e-05, | |
| "loss": 1.3089, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.9361143725395406e-05, | |
| "loss": 1.3078, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.932661095379515e-05, | |
| "loss": 1.2902, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.9292078182194905e-05, | |
| "loss": 1.2991, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.925754541059466e-05, | |
| "loss": 1.2941, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.922301263899441e-05, | |
| "loss": 1.2855, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.9188479867394163e-05, | |
| "loss": 1.3001, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.915394709579391e-05, | |
| "loss": 1.2868, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.911941432419366e-05, | |
| "loss": 1.2751, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.9084881552593415e-05, | |
| "loss": 1.2751, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.905034878099317e-05, | |
| "loss": 1.2725, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.901581600939292e-05, | |
| "loss": 1.2476, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 4.898128323779267e-05, | |
| "loss": 1.2625, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.894675046619242e-05, | |
| "loss": 1.2567, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.891221769459217e-05, | |
| "loss": 1.2474, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.8877684922991925e-05, | |
| "loss": 1.2491, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 4.884315215139167e-05, | |
| "loss": 1.2476, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.8808619379791424e-05, | |
| "loss": 1.2328, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.877408660819118e-05, | |
| "loss": 1.2464, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.873955383659093e-05, | |
| "loss": 1.23, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.870502106499068e-05, | |
| "loss": 1.2462, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.867048829339043e-05, | |
| "loss": 1.2256, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.863595552179018e-05, | |
| "loss": 1.2234, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 4.8601422750189934e-05, | |
| "loss": 1.2223, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.856688997858969e-05, | |
| "loss": 1.2349, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.853235720698944e-05, | |
| "loss": 1.2273, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.8497824435389186e-05, | |
| "loss": 1.2351, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 4.846329166378894e-05, | |
| "loss": 1.2239, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.842875889218869e-05, | |
| "loss": 1.2147, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.8394226120588445e-05, | |
| "loss": 1.2129, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.83596933489882e-05, | |
| "loss": 1.2149, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.8325160577387943e-05, | |
| "loss": 1.2097, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.8290627805787696e-05, | |
| "loss": 1.214, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.825609503418745e-05, | |
| "loss": 1.2035, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 4.82215622625872e-05, | |
| "loss": 1.2006, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.818702949098695e-05, | |
| "loss": 1.1965, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.81524967193867e-05, | |
| "loss": 1.1995, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.8117963947786454e-05, | |
| "loss": 1.1923, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.8083431176186207e-05, | |
| "loss": 1.1872, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.804889840458596e-05, | |
| "loss": 1.1965, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.8014365632985705e-05, | |
| "loss": 1.1886, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.797983286138546e-05, | |
| "loss": 1.1855, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.794530008978521e-05, | |
| "loss": 1.1894, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.7910767318184964e-05, | |
| "loss": 1.1936, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.787623454658471e-05, | |
| "loss": 1.1828, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 4.784170177498446e-05, | |
| "loss": 1.2018, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.780716900338421e-05, | |
| "loss": 1.1821, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.777263623178396e-05, | |
| "loss": 1.1762, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 4.7738103460183714e-05, | |
| "loss": 1.1855, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.770357068858347e-05, | |
| "loss": 1.1738, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.766903791698322e-05, | |
| "loss": 1.1739, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.7634505145382966e-05, | |
| "loss": 1.1756, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.759997237378272e-05, | |
| "loss": 1.1687, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.756543960218247e-05, | |
| "loss": 1.16, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.7530906830582225e-05, | |
| "loss": 1.1658, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.749637405898197e-05, | |
| "loss": 1.1757, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.7461841287381724e-05, | |
| "loss": 1.1742, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.7427308515781476e-05, | |
| "loss": 1.1681, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "eval_validation_accuracy": 0.05685549584758183, | |
| "eval_validation_loss": 1.171875, | |
| "eval_validation_runtime": 3753.295, | |
| "eval_validation_samples_per_second": 0.666, | |
| "eval_validation_steps_per_second": 0.021, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.739277574418123e-05, | |
| "loss": 1.1749, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 4.735824297258098e-05, | |
| "loss": 1.1617, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.732371020098073e-05, | |
| "loss": 1.1696, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.728917742938048e-05, | |
| "loss": 1.1638, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.7254644657780234e-05, | |
| "loss": 1.1634, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.7220111886179987e-05, | |
| "loss": 1.1545, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.718557911457974e-05, | |
| "loss": 1.1667, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.7151046342979485e-05, | |
| "loss": 1.1503, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.711651357137924e-05, | |
| "loss": 1.1633, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 4.708198079977899e-05, | |
| "loss": 1.1653, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.7047448028178744e-05, | |
| "loss": 1.1642, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.70129152565785e-05, | |
| "loss": 1.15, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 4.697838248497824e-05, | |
| "loss": 1.1667, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.6943849713377996e-05, | |
| "loss": 1.1532, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.690931694177775e-05, | |
| "loss": 1.1621, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.68747841701775e-05, | |
| "loss": 1.1509, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.684025139857725e-05, | |
| "loss": 1.1515, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.6805718626977e-05, | |
| "loss": 1.1513, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.677118585537675e-05, | |
| "loss": 1.1538, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 4.6736653083776506e-05, | |
| "loss": 1.146, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.670212031217626e-05, | |
| "loss": 1.1537, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.6667587540576005e-05, | |
| "loss": 1.1524, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.663305476897576e-05, | |
| "loss": 1.1523, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 4.659852199737551e-05, | |
| "loss": 1.15, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.656398922577526e-05, | |
| "loss": 1.1419, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.6529456454175016e-05, | |
| "loss": 1.1525, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.649492368257476e-05, | |
| "loss": 1.1457, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 4.6460390910974515e-05, | |
| "loss": 1.1454, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.642585813937427e-05, | |
| "loss": 1.1284, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.639132536777402e-05, | |
| "loss": 1.1481, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.6356792596173773e-05, | |
| "loss": 1.1484, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.632225982457352e-05, | |
| "loss": 1.1332, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.628772705297327e-05, | |
| "loss": 1.138, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.6253194281373025e-05, | |
| "loss": 1.1369, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.621866150977278e-05, | |
| "loss": 1.1399, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.6184128738172524e-05, | |
| "loss": 1.1221, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.614959596657228e-05, | |
| "loss": 1.1233, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.611506319497203e-05, | |
| "loss": 1.1248, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 4.608053042337178e-05, | |
| "loss": 1.1333, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.6045997651771535e-05, | |
| "loss": 1.1266, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.601146488017128e-05, | |
| "loss": 1.1368, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.5976932108571034e-05, | |
| "loss": 1.1296, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.594239933697079e-05, | |
| "loss": 1.1318, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.590786656537054e-05, | |
| "loss": 1.1272, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.587333379377029e-05, | |
| "loss": 1.1197, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 4.583880102217004e-05, | |
| "loss": 1.1271, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.580426825056979e-05, | |
| "loss": 1.1193, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.5769735478969544e-05, | |
| "loss": 1.1251, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 4.57352027073693e-05, | |
| "loss": 1.123, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.570066993576905e-05, | |
| "loss": 1.1141, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.5666137164168796e-05, | |
| "loss": 1.1301, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.563160439256855e-05, | |
| "loss": 1.1106, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.55970716209683e-05, | |
| "loss": 1.1258, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.5562538849368055e-05, | |
| "loss": 1.1198, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.552800607776781e-05, | |
| "loss": 1.1157, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.5493473306167554e-05, | |
| "loss": 1.1277, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.5458940534567306e-05, | |
| "loss": 1.1169, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.542440776296706e-05, | |
| "loss": 1.1062, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.538987499136681e-05, | |
| "loss": 1.1187, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.535534221976656e-05, | |
| "loss": 1.1095, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.532080944816631e-05, | |
| "loss": 1.1235, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.5286276676566064e-05, | |
| "loss": 1.1165, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.5251743904965817e-05, | |
| "loss": 1.117, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.521721113336557e-05, | |
| "loss": 1.1073, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.5182678361765315e-05, | |
| "loss": 1.1156, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.514814559016507e-05, | |
| "loss": 1.1034, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.511361281856482e-05, | |
| "loss": 1.1127, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.5079080046964574e-05, | |
| "loss": 1.1214, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.504454727536433e-05, | |
| "loss": 1.116, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.501001450376407e-05, | |
| "loss": 1.1178, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.4975481732163826e-05, | |
| "loss": 1.1048, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.494094896056358e-05, | |
| "loss": 1.1026, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.490641618896333e-05, | |
| "loss": 1.1151, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.4871883417363084e-05, | |
| "loss": 1.0999, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.483735064576283e-05, | |
| "loss": 1.1257, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "eval_validation_accuracy": 0.0578276502198339, | |
| "eval_validation_loss": 1.109375, | |
| "eval_validation_runtime": 3693.3851, | |
| "eval_validation_samples_per_second": 0.677, | |
| "eval_validation_steps_per_second": 0.021, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.480281787416258e-05, | |
| "loss": 1.1124, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.4768285102562336e-05, | |
| "loss": 1.1108, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.473375233096209e-05, | |
| "loss": 1.0994, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.4699219559361835e-05, | |
| "loss": 1.11, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.466468678776159e-05, | |
| "loss": 1.101, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.463015401616134e-05, | |
| "loss": 1.0904, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.459562124456109e-05, | |
| "loss": 1.1135, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.4561088472960846e-05, | |
| "loss": 1.0917, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.452655570136059e-05, | |
| "loss": 1.109, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.4492022929760345e-05, | |
| "loss": 1.0901, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.44574901581601e-05, | |
| "loss": 1.1037, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.442295738655985e-05, | |
| "loss": 1.1057, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.4388424614959603e-05, | |
| "loss": 1.1004, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.435389184335935e-05, | |
| "loss": 1.1008, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.43193590717591e-05, | |
| "loss": 1.1036, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.4284826300158855e-05, | |
| "loss": 1.1124, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.425029352855861e-05, | |
| "loss": 1.1096, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.421576075695836e-05, | |
| "loss": 1.1069, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.418122798535811e-05, | |
| "loss": 1.0886, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.414669521375786e-05, | |
| "loss": 1.0998, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.411216244215761e-05, | |
| "loss": 1.0858, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.4077629670557365e-05, | |
| "loss": 1.0987, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.404309689895711e-05, | |
| "loss": 1.0952, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.4010290765936875e-05, | |
| "loss": 1.1043, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.397575799433663e-05, | |
| "loss": 1.0918, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.3941225222736374e-05, | |
| "loss": 1.0952, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.3906692451136127e-05, | |
| "loss": 1.0997, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.387215967953588e-05, | |
| "loss": 1.1022, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.383762690793563e-05, | |
| "loss": 1.0985, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.3803094136335385e-05, | |
| "loss": 1.0928, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.376856136473513e-05, | |
| "loss": 1.0793, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.3734028593134884e-05, | |
| "loss": 1.0927, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.369949582153464e-05, | |
| "loss": 1.0959, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.366496304993439e-05, | |
| "loss": 1.0914, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.3630430278334136e-05, | |
| "loss": 1.0874, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.359589750673389e-05, | |
| "loss": 1.0954, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.356136473513364e-05, | |
| "loss": 1.0906, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.3526831963533394e-05, | |
| "loss": 1.0896, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.349229919193315e-05, | |
| "loss": 1.096, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.345776642033289e-05, | |
| "loss": 1.0946, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.3423233648732646e-05, | |
| "loss": 1.0868, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.33887008771324e-05, | |
| "loss": 1.0914, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.335416810553215e-05, | |
| "loss": 1.0836, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.3319635333931904e-05, | |
| "loss": 1.0865, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.328510256233165e-05, | |
| "loss": 1.0922, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.32505697907314e-05, | |
| "loss": 1.0891, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.3216037019131156e-05, | |
| "loss": 1.0836, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.318150424753091e-05, | |
| "loss": 1.0914, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.314697147593066e-05, | |
| "loss": 1.0772, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.311243870433041e-05, | |
| "loss": 1.0849, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.307790593273016e-05, | |
| "loss": 1.0747, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.3043373161129913e-05, | |
| "loss": 1.0904, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.3008840389529666e-05, | |
| "loss": 1.0839, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.297430761792942e-05, | |
| "loss": 1.0799, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.2939774846329165e-05, | |
| "loss": 1.0787, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.290524207472892e-05, | |
| "loss": 1.0659, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.287070930312867e-05, | |
| "loss": 1.0821, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.2836176531528424e-05, | |
| "loss": 1.0857, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.280164375992817e-05, | |
| "loss": 1.0834, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.276711098832792e-05, | |
| "loss": 1.0793, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.2732578216727675e-05, | |
| "loss": 1.0692, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.269804544512743e-05, | |
| "loss": 1.0762, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.266351267352718e-05, | |
| "loss": 1.0859, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.262897990192693e-05, | |
| "loss": 1.0827, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.259444713032668e-05, | |
| "loss": 1.0787, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.255991435872643e-05, | |
| "loss": 1.0711, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.2525381587126186e-05, | |
| "loss": 1.0852, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.249084881552594e-05, | |
| "loss": 1.0719, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.2456316043925684e-05, | |
| "loss": 1.0877, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.242178327232544e-05, | |
| "loss": 1.0856, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.238725050072519e-05, | |
| "loss": 1.0704, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.235271772912494e-05, | |
| "loss": 1.0776, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.2318184957524696e-05, | |
| "loss": 1.0751, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.228365218592444e-05, | |
| "loss": 1.0801, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.2249119414324195e-05, | |
| "loss": 1.0891, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "eval_validation_accuracy": 0.0582704445530044, | |
| "eval_validation_loss": 1.0771484375, | |
| "eval_validation_runtime": 3796.035, | |
| "eval_validation_samples_per_second": 0.659, | |
| "eval_validation_steps_per_second": 0.021, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.221458664272395e-05, | |
| "loss": 1.0832, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.21800538711237e-05, | |
| "loss": 1.076, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.2145521099523446e-05, | |
| "loss": 1.0764, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.21109883279232e-05, | |
| "loss": 1.0848, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.207645555632295e-05, | |
| "loss": 1.0813, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.2041922784722705e-05, | |
| "loss": 1.081, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.200739001312246e-05, | |
| "loss": 1.0664, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.1972857241522204e-05, | |
| "loss": 1.074, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.1938324469921957e-05, | |
| "loss": 1.069, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.190379169832171e-05, | |
| "loss": 1.0788, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.186925892672146e-05, | |
| "loss": 1.0631, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.1836452793701226e-05, | |
| "loss": 1.0728, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.180192002210098e-05, | |
| "loss": 1.0665, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.176911388908074e-05, | |
| "loss": 1.0753, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.1734581117480495e-05, | |
| "loss": 1.0741, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.170004834588024e-05, | |
| "loss": 1.071, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.1665515574279994e-05, | |
| "loss": 1.0826, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.1630982802679747e-05, | |
| "loss": 1.067, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.15964500310795e-05, | |
| "loss": 1.0671, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.156191725947925e-05, | |
| "loss": 1.0666, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.1527384487879e-05, | |
| "loss": 1.0749, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.149285171627875e-05, | |
| "loss": 1.0772, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.1458318944678504e-05, | |
| "loss": 1.0762, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.142378617307826e-05, | |
| "loss": 1.0741, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.138925340147801e-05, | |
| "loss": 1.0701, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.1354720629877756e-05, | |
| "loss": 1.0724, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.132018785827751e-05, | |
| "loss": 1.0651, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.128565508667726e-05, | |
| "loss": 1.0653, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.1251122315077014e-05, | |
| "loss": 1.0854, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.121658954347676e-05, | |
| "loss": 1.064, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.118205677187651e-05, | |
| "loss": 1.0818, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.1147524000276266e-05, | |
| "loss": 1.0686, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.111299122867602e-05, | |
| "loss": 1.0708, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.107845845707577e-05, | |
| "loss": 1.0551, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.104392568547552e-05, | |
| "loss": 1.0667, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.100939291387527e-05, | |
| "loss": 1.0588, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.097486014227502e-05, | |
| "loss": 1.0591, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.0940327370674776e-05, | |
| "loss": 1.085, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.090579459907453e-05, | |
| "loss": 1.0772, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.0871261827474275e-05, | |
| "loss": 1.0565, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.083672905587403e-05, | |
| "loss": 1.0626, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.080219628427378e-05, | |
| "loss": 1.0589, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.0767663512673533e-05, | |
| "loss": 1.0693, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.0733130741073286e-05, | |
| "loss": 1.0834, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.069859796947303e-05, | |
| "loss": 1.0668, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.066406519787278e-05, | |
| "loss": 1.0722, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.062953242627253e-05, | |
| "loss": 1.0591, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.0594999654672284e-05, | |
| "loss": 1.0579, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.056046688307204e-05, | |
| "loss": 1.0488, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.052593411147178e-05, | |
| "loss": 1.0575, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.0491401339871536e-05, | |
| "loss": 1.0654, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.045686856827129e-05, | |
| "loss": 1.0793, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.042233579667104e-05, | |
| "loss": 1.0717, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.0387803025070794e-05, | |
| "loss": 1.0772, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.035327025347054e-05, | |
| "loss": 1.0641, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.031873748187029e-05, | |
| "loss": 1.0649, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.0284204710270046e-05, | |
| "loss": 1.0728, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.02496719386698e-05, | |
| "loss": 1.0697, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.021513916706955e-05, | |
| "loss": 1.0575, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.01806063954693e-05, | |
| "loss": 1.0569, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.014607362386905e-05, | |
| "loss": 1.0655, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.01115408522688e-05, | |
| "loss": 1.0669, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.0077008080668556e-05, | |
| "loss": 1.0599, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.004247530906831e-05, | |
| "loss": 1.0731, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.0007942537468055e-05, | |
| "loss": 1.0596, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 3.997340976586781e-05, | |
| "loss": 1.0625, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.993887699426756e-05, | |
| "loss": 1.0601, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.9904344222667314e-05, | |
| "loss": 1.0605, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.986981145106706e-05, | |
| "loss": 1.0535, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 3.983527867946681e-05, | |
| "loss": 1.0612, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.9800745907866565e-05, | |
| "loss": 1.0652, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.976621313626632e-05, | |
| "loss": 1.0616, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 3.973168036466607e-05, | |
| "loss": 1.0546, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.969714759306582e-05, | |
| "loss": 1.0518, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.966434146004559e-05, | |
| "loss": 1.0609, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "eval_validation_accuracy": 0.05839198827552516, | |
| "eval_validation_loss": 1.0556640625, | |
| "eval_validation_runtime": 3758.5842, | |
| "eval_validation_samples_per_second": 0.665, | |
| "eval_validation_steps_per_second": 0.021, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.962980868844533e-05, | |
| "loss": 1.0554, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 3.9595275916845086e-05, | |
| "loss": 1.0507, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.956074314524484e-05, | |
| "loss": 1.0517, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.952621037364459e-05, | |
| "loss": 1.0482, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 3.9491677602044345e-05, | |
| "loss": 1.0634, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.945714483044409e-05, | |
| "loss": 1.0629, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.9422612058843843e-05, | |
| "loss": 1.0474, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.9388079287243596e-05, | |
| "loss": 1.0513, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 3.935527315422336e-05, | |
| "loss": 1.0527, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.932074038262311e-05, | |
| "loss": 1.0482, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.9286207611022865e-05, | |
| "loss": 1.069, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.925167483942261e-05, | |
| "loss": 1.0485, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 3.9217142067822364e-05, | |
| "loss": 1.0525, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.918260929622212e-05, | |
| "loss": 1.0424, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.914807652462187e-05, | |
| "loss": 1.0501, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 3.911354375302162e-05, | |
| "loss": 1.048, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.907901098142137e-05, | |
| "loss": 1.0454, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.904447820982112e-05, | |
| "loss": 1.0691, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.9009945438220875e-05, | |
| "loss": 1.0534, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 3.897541266662063e-05, | |
| "loss": 1.0413, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.8940879895020373e-05, | |
| "loss": 1.0564, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.8906347123420126e-05, | |
| "loss": 1.0589, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.887181435181988e-05, | |
| "loss": 1.0495, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 3.883900821879964e-05, | |
| "loss": 1.0499, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.8804475447199395e-05, | |
| "loss": 1.0551, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.876994267559915e-05, | |
| "loss": 1.0399, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.87354099039989e-05, | |
| "loss": 1.0541, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.870087713239865e-05, | |
| "loss": 1.0507, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.86663443607984e-05, | |
| "loss": 1.0489, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.863181158919815e-05, | |
| "loss": 1.0456, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 3.8597278817597906e-05, | |
| "loss": 1.0476, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.856274604599766e-05, | |
| "loss": 1.0498, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.8528213274397404e-05, | |
| "loss": 1.0444, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 3.849368050279716e-05, | |
| "loss": 1.059, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.845914773119691e-05, | |
| "loss": 1.0575, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.842461495959666e-05, | |
| "loss": 1.0364, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.839008218799641e-05, | |
| "loss": 1.0455, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 3.835554941639616e-05, | |
| "loss": 1.0484, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.8321016644795915e-05, | |
| "loss": 1.0263, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.828648387319567e-05, | |
| "loss": 1.048, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.825195110159542e-05, | |
| "loss": 1.0526, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 3.8217418329995166e-05, | |
| "loss": 1.0433, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.818288555839492e-05, | |
| "loss": 1.0491, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.814835278679467e-05, | |
| "loss": 1.0437, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 3.8113820015194425e-05, | |
| "loss": 1.0498, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.807928724359418e-05, | |
| "loss": 1.0457, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.8044754471993924e-05, | |
| "loss": 1.0481, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.8010221700393677e-05, | |
| "loss": 1.0372, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.797568892879343e-05, | |
| "loss": 1.056, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.794115615719318e-05, | |
| "loss": 1.0413, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.7906623385592935e-05, | |
| "loss": 1.0426, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.787209061399268e-05, | |
| "loss": 1.043, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.7837557842392434e-05, | |
| "loss": 1.0504, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.780302507079219e-05, | |
| "loss": 1.0467, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.776849229919194e-05, | |
| "loss": 1.047, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.7733959527591686e-05, | |
| "loss": 1.0384, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.769942675599143e-05, | |
| "loss": 1.0451, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.7664893984391185e-05, | |
| "loss": 1.0366, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.763036121279094e-05, | |
| "loss": 1.0403, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.759582844119069e-05, | |
| "loss": 1.0315, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.756129566959044e-05, | |
| "loss": 1.0303, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.752676289799019e-05, | |
| "loss": 1.0395, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.749223012638994e-05, | |
| "loss": 0.985, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.7457697354789695e-05, | |
| "loss": 0.9691, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.742316458318945e-05, | |
| "loss": 0.9904, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.73886318115892e-05, | |
| "loss": 0.973, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.7354099039988946e-05, | |
| "loss": 0.9727, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.73195662683887e-05, | |
| "loss": 0.9801, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.728503349678845e-05, | |
| "loss": 0.9896, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.7250500725188205e-05, | |
| "loss": 0.9718, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.721596795358796e-05, | |
| "loss": 0.9824, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.7181435181987704e-05, | |
| "loss": 0.9742, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.714690241038746e-05, | |
| "loss": 0.9716, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.711236963878721e-05, | |
| "loss": 0.9822, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.707783686718696e-05, | |
| "loss": 0.9863, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "eval_validation_accuracy": 0.05860850024425989, | |
| "eval_validation_loss": 1.0419921875, | |
| "eval_validation_runtime": 3723.0597, | |
| "eval_validation_samples_per_second": 0.671, | |
| "eval_validation_steps_per_second": 0.021, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.704330409558671e-05, | |
| "loss": 0.979, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.700877132398646e-05, | |
| "loss": 0.9744, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.6974238552386214e-05, | |
| "loss": 0.9763, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.693970578078597e-05, | |
| "loss": 0.9725, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.690517300918572e-05, | |
| "loss": 0.9691, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.6870640237585466e-05, | |
| "loss": 0.9741, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.683610746598522e-05, | |
| "loss": 0.9836, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.680157469438497e-05, | |
| "loss": 0.9823, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.6767041922784724e-05, | |
| "loss": 0.9781, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.673250915118448e-05, | |
| "loss": 0.9754, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.669797637958422e-05, | |
| "loss": 0.9769, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.6663443607983976e-05, | |
| "loss": 0.9831, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.662891083638373e-05, | |
| "loss": 0.9825, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.659437806478348e-05, | |
| "loss": 0.9838, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.6559845293183234e-05, | |
| "loss": 0.97, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.652531252158298e-05, | |
| "loss": 0.9849, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.649077974998273e-05, | |
| "loss": 0.9854, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.6456246978382486e-05, | |
| "loss": 0.9765, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.642171420678224e-05, | |
| "loss": 0.9758, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.6387181435181985e-05, | |
| "loss": 0.9696, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.635264866358174e-05, | |
| "loss": 0.9751, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.631811589198149e-05, | |
| "loss": 0.979, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.6283583120381244e-05, | |
| "loss": 0.9609, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.6249050348780996e-05, | |
| "loss": 0.9798, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.621451757718074e-05, | |
| "loss": 0.9627, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.6179984805580495e-05, | |
| "loss": 0.981, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.614545203398025e-05, | |
| "loss": 0.9759, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.611091926238e-05, | |
| "loss": 0.9782, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.6076386490779754e-05, | |
| "loss": 0.9778, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.60418537191795e-05, | |
| "loss": 0.9752, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.600732094757925e-05, | |
| "loss": 0.9852, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.5972788175979005e-05, | |
| "loss": 0.9776, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.593825540437876e-05, | |
| "loss": 0.9742, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.590372263277851e-05, | |
| "loss": 0.9716, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.586918986117826e-05, | |
| "loss": 0.9744, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.583465708957801e-05, | |
| "loss": 0.9738, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.580012431797776e-05, | |
| "loss": 0.9851, | |
| "step": 8240 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.5765591546377516e-05, | |
| "loss": 0.9794, | |
| "step": 8260 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.573105877477727e-05, | |
| "loss": 0.9785, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.5696526003177015e-05, | |
| "loss": 0.978, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.566199323157677e-05, | |
| "loss": 0.9681, | |
| "step": 8320 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.562746045997652e-05, | |
| "loss": 0.9849, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.559292768837627e-05, | |
| "loss": 0.9799, | |
| "step": 8360 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.555839491677602e-05, | |
| "loss": 0.9694, | |
| "step": 8380 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.552386214517577e-05, | |
| "loss": 0.9724, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.5489329373575525e-05, | |
| "loss": 0.9705, | |
| "step": 8420 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.545479660197528e-05, | |
| "loss": 0.9846, | |
| "step": 8440 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.542026383037503e-05, | |
| "loss": 0.9729, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.5385731058774776e-05, | |
| "loss": 0.9741, | |
| "step": 8480 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.535119828717453e-05, | |
| "loss": 0.9801, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.531666551557428e-05, | |
| "loss": 0.9863, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.5282132743974035e-05, | |
| "loss": 0.9856, | |
| "step": 8540 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.524759997237379e-05, | |
| "loss": 0.9845, | |
| "step": 8560 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.5213067200773534e-05, | |
| "loss": 0.9791, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.517853442917329e-05, | |
| "loss": 0.9719, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.514400165757304e-05, | |
| "loss": 0.9714, | |
| "step": 8620 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.510946888597279e-05, | |
| "loss": 0.9802, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.5074936114372545e-05, | |
| "loss": 0.9823, | |
| "step": 8660 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.504040334277229e-05, | |
| "loss": 0.9808, | |
| "step": 8680 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.5005870571172044e-05, | |
| "loss": 0.9788, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.49713377995718e-05, | |
| "loss": 0.9805, | |
| "step": 8720 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.493680502797155e-05, | |
| "loss": 0.9724, | |
| "step": 8740 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.4902272256371296e-05, | |
| "loss": 0.9887, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.486773948477105e-05, | |
| "loss": 0.9863, | |
| "step": 8780 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.48332067131708e-05, | |
| "loss": 0.9789, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.4798673941570554e-05, | |
| "loss": 0.9738, | |
| "step": 8820 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.476414116997031e-05, | |
| "loss": 0.9881, | |
| "step": 8840 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.472960839837005e-05, | |
| "loss": 0.9718, | |
| "step": 8860 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.4695075626769806e-05, | |
| "loss": 0.9842, | |
| "step": 8880 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.466054285516956e-05, | |
| "loss": 0.9775, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.462601008356931e-05, | |
| "loss": 0.9701, | |
| "step": 8920 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.4591477311969064e-05, | |
| "loss": 0.9826, | |
| "step": 8940 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.455694454036881e-05, | |
| "loss": 0.9734, | |
| "step": 8960 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.452241176876856e-05, | |
| "loss": 0.9713, | |
| "step": 8980 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.4487878997168316e-05, | |
| "loss": 0.9813, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "eval_validation_accuracy": 0.05870815828041036, | |
| "eval_validation_loss": 1.03515625, | |
| "eval_validation_runtime": 3747.9661, | |
| "eval_validation_samples_per_second": 0.667, | |
| "eval_validation_steps_per_second": 0.021, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.445334622556807e-05, | |
| "loss": 0.9817, | |
| "step": 9020 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.441881345396782e-05, | |
| "loss": 0.9692, | |
| "step": 9040 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.438428068236757e-05, | |
| "loss": 0.9694, | |
| "step": 9060 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.434974791076732e-05, | |
| "loss": 0.9659, | |
| "step": 9080 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.4315215139167074e-05, | |
| "loss": 0.9697, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.4280682367566826e-05, | |
| "loss": 0.9747, | |
| "step": 9120 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.424614959596657e-05, | |
| "loss": 0.98, | |
| "step": 9140 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.4211616824366325e-05, | |
| "loss": 0.9731, | |
| "step": 9160 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.417708405276608e-05, | |
| "loss": 0.9645, | |
| "step": 9180 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.414255128116583e-05, | |
| "loss": 0.986, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.4108018509565584e-05, | |
| "loss": 0.9826, | |
| "step": 9220 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.407348573796533e-05, | |
| "loss": 0.9805, | |
| "step": 9240 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.403895296636508e-05, | |
| "loss": 0.9782, | |
| "step": 9260 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.4004420194764835e-05, | |
| "loss": 0.9863, | |
| "step": 9280 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.396988742316459e-05, | |
| "loss": 0.9724, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.393535465156434e-05, | |
| "loss": 0.9727, | |
| "step": 9320 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.390082187996409e-05, | |
| "loss": 0.9743, | |
| "step": 9340 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.386628910836384e-05, | |
| "loss": 0.9782, | |
| "step": 9360 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.383175633676359e-05, | |
| "loss": 0.9744, | |
| "step": 9380 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.3797223565163346e-05, | |
| "loss": 0.9681, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.37626907935631e-05, | |
| "loss": 0.9878, | |
| "step": 9420 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.3728158021962845e-05, | |
| "loss": 0.9784, | |
| "step": 9440 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.36936252503626e-05, | |
| "loss": 0.9665, | |
| "step": 9460 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.365909247876235e-05, | |
| "loss": 0.9742, | |
| "step": 9480 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.36245597071621e-05, | |
| "loss": 0.965, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.359002693556185e-05, | |
| "loss": 0.9742, | |
| "step": 9520 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.35554941639616e-05, | |
| "loss": 0.9744, | |
| "step": 9540 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.3520961392361355e-05, | |
| "loss": 0.9817, | |
| "step": 9560 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.348642862076111e-05, | |
| "loss": 0.9752, | |
| "step": 9580 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.3451895849160854e-05, | |
| "loss": 0.9851, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.3417363077560606e-05, | |
| "loss": 0.9791, | |
| "step": 9620 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.338283030596035e-05, | |
| "loss": 0.9719, | |
| "step": 9640 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.3348297534360105e-05, | |
| "loss": 0.9627, | |
| "step": 9660 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.331376476275986e-05, | |
| "loss": 0.9719, | |
| "step": 9680 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.327923199115961e-05, | |
| "loss": 0.9627, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.3244699219559364e-05, | |
| "loss": 0.9758, | |
| "step": 9720 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.321016644795911e-05, | |
| "loss": 0.9776, | |
| "step": 9740 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.317563367635886e-05, | |
| "loss": 0.9581, | |
| "step": 9760 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.3141100904758616e-05, | |
| "loss": 0.9667, | |
| "step": 9780 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.310656813315837e-05, | |
| "loss": 0.9544, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.307203536155812e-05, | |
| "loss": 0.9753, | |
| "step": 9820 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.303750258995787e-05, | |
| "loss": 0.9687, | |
| "step": 9840 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.300296981835762e-05, | |
| "loss": 0.9813, | |
| "step": 9860 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.296843704675737e-05, | |
| "loss": 0.9627, | |
| "step": 9880 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.2933904275157126e-05, | |
| "loss": 0.9754, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.289937150355687e-05, | |
| "loss": 0.9763, | |
| "step": 9920 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.2864838731956625e-05, | |
| "loss": 0.9742, | |
| "step": 9940 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.283030596035638e-05, | |
| "loss": 0.9606, | |
| "step": 9960 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.279577318875613e-05, | |
| "loss": 0.9692, | |
| "step": 9980 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.276124041715588e-05, | |
| "loss": 0.9624, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.272670764555563e-05, | |
| "loss": 0.9659, | |
| "step": 10020 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.269217487395538e-05, | |
| "loss": 0.9746, | |
| "step": 10040 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.2657642102355135e-05, | |
| "loss": 0.967, | |
| "step": 10060 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.262310933075489e-05, | |
| "loss": 0.974, | |
| "step": 10080 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.258857655915464e-05, | |
| "loss": 0.9745, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.2554043787554387e-05, | |
| "loss": 0.9803, | |
| "step": 10120 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.251951101595414e-05, | |
| "loss": 0.9752, | |
| "step": 10140 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.248497824435389e-05, | |
| "loss": 0.9692, | |
| "step": 10160 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.2450445472753645e-05, | |
| "loss": 0.972, | |
| "step": 10180 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.24159127011534e-05, | |
| "loss": 0.9833, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.2381379929553144e-05, | |
| "loss": 0.9744, | |
| "step": 10220 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.23468471579529e-05, | |
| "loss": 0.9632, | |
| "step": 10240 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.231231438635265e-05, | |
| "loss": 0.9722, | |
| "step": 10260 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.22777816147524e-05, | |
| "loss": 0.9773, | |
| "step": 10280 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.224324884315215e-05, | |
| "loss": 0.9657, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.22087160715519e-05, | |
| "loss": 0.9763, | |
| "step": 10320 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.2174183299951654e-05, | |
| "loss": 0.9798, | |
| "step": 10340 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.213965052835141e-05, | |
| "loss": 0.9653, | |
| "step": 10360 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.210511775675116e-05, | |
| "loss": 0.9777, | |
| "step": 10380 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.2070584985150906e-05, | |
| "loss": 0.9742, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.203605221355066e-05, | |
| "loss": 0.9651, | |
| "step": 10420 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.200151944195041e-05, | |
| "loss": 0.9809, | |
| "step": 10440 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.1966986670350164e-05, | |
| "loss": 0.977, | |
| "step": 10460 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.193245389874992e-05, | |
| "loss": 0.9675, | |
| "step": 10480 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.189792112714966e-05, | |
| "loss": 0.968, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "eval_validation_accuracy": 0.058771861260381045, | |
| "eval_validation_loss": 1.0263671875, | |
| "eval_validation_runtime": 3706.0967, | |
| "eval_validation_samples_per_second": 0.675, | |
| "eval_validation_steps_per_second": 0.021, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.3734290843806105e-05, | |
| "loss": 0.97, | |
| "step": 10520 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.366523960778898e-05, | |
| "loss": 0.9648, | |
| "step": 10540 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.3596188371771857e-05, | |
| "loss": 0.9569, | |
| "step": 10560 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.352713713575473e-05, | |
| "loss": 0.9731, | |
| "step": 10580 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.3458085899737605e-05, | |
| "loss": 0.9581, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.3392487225521338e-05, | |
| "loss": 0.9711, | |
| "step": 10620 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.3323435989504214e-05, | |
| "loss": 0.9612, | |
| "step": 10640 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.3254384753487087e-05, | |
| "loss": 0.9696, | |
| "step": 10660 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.3185333517469964e-05, | |
| "loss": 0.9738, | |
| "step": 10680 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.3116282281452839e-05, | |
| "loss": 0.9559, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.3047231045435716e-05, | |
| "loss": 0.962, | |
| "step": 10720 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.2978179809418589e-05, | |
| "loss": 0.9499, | |
| "step": 10740 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.291258113520232e-05, | |
| "loss": 0.9599, | |
| "step": 10760 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.2843529899185197e-05, | |
| "loss": 0.9667, | |
| "step": 10780 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.2777931224968928e-05, | |
| "loss": 0.9513, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.2708879988951802e-05, | |
| "loss": 0.9456, | |
| "step": 10820 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.2639828752934679e-05, | |
| "loss": 0.9569, | |
| "step": 10840 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.2570777516917554e-05, | |
| "loss": 0.9531, | |
| "step": 10860 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.2501726280900427e-05, | |
| "loss": 0.9626, | |
| "step": 10880 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.2432675044883304e-05, | |
| "loss": 0.9502, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.2363623808866179e-05, | |
| "loss": 0.9624, | |
| "step": 10920 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.2294572572849054e-05, | |
| "loss": 0.9558, | |
| "step": 10940 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.2225521336831929e-05, | |
| "loss": 0.9466, | |
| "step": 10960 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.2156470100814806e-05, | |
| "loss": 0.9497, | |
| "step": 10980 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.2087418864797681e-05, | |
| "loss": 0.9523, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.2018367628780556e-05, | |
| "loss": 0.951, | |
| "step": 11020 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.1949316392763431e-05, | |
| "loss": 0.9636, | |
| "step": 11040 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.1880265156746306e-05, | |
| "loss": 0.9631, | |
| "step": 11060 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.1811213920729181e-05, | |
| "loss": 0.9474, | |
| "step": 11080 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.1742162684712058e-05, | |
| "loss": 0.9698, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.1673111448694931e-05, | |
| "loss": 0.9616, | |
| "step": 11120 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.1604060212677806e-05, | |
| "loss": 0.96, | |
| "step": 11140 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.1535008976660683e-05, | |
| "loss": 0.9632, | |
| "step": 11160 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.1465957740643558e-05, | |
| "loss": 0.9508, | |
| "step": 11180 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.1396906504626433e-05, | |
| "loss": 0.9587, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.1327855268609308e-05, | |
| "loss": 0.9466, | |
| "step": 11220 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.1258804032592184e-05, | |
| "loss": 0.9518, | |
| "step": 11240 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.1189752796575059e-05, | |
| "loss": 0.95, | |
| "step": 11260 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.1120701560557935e-05, | |
| "loss": 0.9504, | |
| "step": 11280 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.105165032454081e-05, | |
| "loss": 0.958, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.0982599088523684e-05, | |
| "loss": 0.9506, | |
| "step": 11320 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.091354785250656e-05, | |
| "loss": 0.9422, | |
| "step": 11340 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.0844496616489436e-05, | |
| "loss": 0.9462, | |
| "step": 11360 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.077544538047231e-05, | |
| "loss": 0.9432, | |
| "step": 11380 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.0706394144455187e-05, | |
| "loss": 0.9396, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.0637342908438061e-05, | |
| "loss": 0.9434, | |
| "step": 11420 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.0568291672420936e-05, | |
| "loss": 0.954, | |
| "step": 11440 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.0499240436403813e-05, | |
| "loss": 0.9513, | |
| "step": 11460 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.0430189200386688e-05, | |
| "loss": 0.9538, | |
| "step": 11480 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.0361137964369563e-05, | |
| "loss": 0.9518, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.0292086728352438e-05, | |
| "loss": 0.9603, | |
| "step": 11520 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.0223035492335313e-05, | |
| "loss": 0.9567, | |
| "step": 11540 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.0153984256318188e-05, | |
| "loss": 0.9569, | |
| "step": 11560 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.0084933020301065e-05, | |
| "loss": 0.9496, | |
| "step": 11580 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.001588178428394e-05, | |
| "loss": 0.9549, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.946830548266813e-06, | |
| "loss": 0.9547, | |
| "step": 11620 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.87777931224969e-06, | |
| "loss": 0.9597, | |
| "step": 11640 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.808728076232565e-06, | |
| "loss": 0.9566, | |
| "step": 11660 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 9.73967684021544e-06, | |
| "loss": 0.9474, | |
| "step": 11680 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.670625604198317e-06, | |
| "loss": 0.9548, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.60157436818119e-06, | |
| "loss": 0.9529, | |
| "step": 11720 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.532523132164065e-06, | |
| "loss": 0.9418, | |
| "step": 11740 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.463471896146942e-06, | |
| "loss": 0.9463, | |
| "step": 11760 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.394420660129817e-06, | |
| "loss": 0.9557, | |
| "step": 11780 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.325369424112692e-06, | |
| "loss": 0.9495, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 9.256318188095567e-06, | |
| "loss": 0.9496, | |
| "step": 11820 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.187266952078443e-06, | |
| "loss": 0.955, | |
| "step": 11840 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.118215716061318e-06, | |
| "loss": 0.9505, | |
| "step": 11860 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.049164480044194e-06, | |
| "loss": 0.9545, | |
| "step": 11880 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 8.98011324402707e-06, | |
| "loss": 0.9602, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.911062008009943e-06, | |
| "loss": 0.9466, | |
| "step": 11920 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.84201077199282e-06, | |
| "loss": 0.9423, | |
| "step": 11940 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.772959535975695e-06, | |
| "loss": 0.9469, | |
| "step": 11960 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 8.70390829995857e-06, | |
| "loss": 0.9347, | |
| "step": 11980 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.634857063941445e-06, | |
| "loss": 0.9444, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "eval_validation_accuracy": 0.059099560332193456, | |
| "eval_validation_loss": 1.0048828125, | |
| "eval_validation_runtime": 3761.2942, | |
| "eval_validation_samples_per_second": 0.665, | |
| "eval_validation_steps_per_second": 0.021, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.56580582792432e-06, | |
| "loss": 0.9547, | |
| "step": 12020 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 8.496754591907195e-06, | |
| "loss": 0.9516, | |
| "step": 12040 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.42770335589007e-06, | |
| "loss": 0.9476, | |
| "step": 12060 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.358652119872947e-06, | |
| "loss": 0.9455, | |
| "step": 12080 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.289600883855822e-06, | |
| "loss": 0.9453, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 8.220549647838697e-06, | |
| "loss": 0.9336, | |
| "step": 12120 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.151498411821572e-06, | |
| "loss": 0.9553, | |
| "step": 12140 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.082447175804447e-06, | |
| "loss": 0.9419, | |
| "step": 12160 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.013395939787322e-06, | |
| "loss": 0.938, | |
| "step": 12180 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 7.944344703770199e-06, | |
| "loss": 0.9498, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 7.875293467753072e-06, | |
| "loss": 0.9559, | |
| "step": 12220 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 7.806242231735947e-06, | |
| "loss": 0.9426, | |
| "step": 12240 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 7.737190995718824e-06, | |
| "loss": 0.9384, | |
| "step": 12260 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.6681397597017e-06, | |
| "loss": 0.9393, | |
| "step": 12280 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.599088523684575e-06, | |
| "loss": 0.947, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.530037287667449e-06, | |
| "loss": 0.9624, | |
| "step": 12320 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 7.4609860516503245e-06, | |
| "loss": 0.9476, | |
| "step": 12340 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 7.3919348156332e-06, | |
| "loss": 0.9529, | |
| "step": 12360 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 7.3228835796160755e-06, | |
| "loss": 0.9414, | |
| "step": 12380 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 7.253832343598951e-06, | |
| "loss": 0.9393, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 7.184781107581826e-06, | |
| "loss": 0.9482, | |
| "step": 12420 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 7.115729871564701e-06, | |
| "loss": 0.9416, | |
| "step": 12440 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 7.046678635547577e-06, | |
| "loss": 0.9353, | |
| "step": 12460 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 6.9776273995304526e-06, | |
| "loss": 0.941, | |
| "step": 12480 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 6.908576163513328e-06, | |
| "loss": 0.9452, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 6.839524927496202e-06, | |
| "loss": 0.9494, | |
| "step": 12520 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 6.770473691479078e-06, | |
| "loss": 0.9493, | |
| "step": 12540 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 6.701422455461953e-06, | |
| "loss": 0.943, | |
| "step": 12560 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 6.632371219444829e-06, | |
| "loss": 0.9486, | |
| "step": 12580 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 6.563319983427704e-06, | |
| "loss": 0.9506, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 6.494268747410578e-06, | |
| "loss": 0.9474, | |
| "step": 12620 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 6.425217511393454e-06, | |
| "loss": 0.9442, | |
| "step": 12640 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 6.35616627537633e-06, | |
| "loss": 0.9527, | |
| "step": 12660 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 6.287115039359205e-06, | |
| "loss": 0.9332, | |
| "step": 12680 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 6.21806380334208e-06, | |
| "loss": 0.943, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 6.149012567324956e-06, | |
| "loss": 0.9441, | |
| "step": 12720 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 6.07996133130783e-06, | |
| "loss": 0.9359, | |
| "step": 12740 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 6.010910095290706e-06, | |
| "loss": 0.9368, | |
| "step": 12760 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 5.941858859273581e-06, | |
| "loss": 0.9466, | |
| "step": 12780 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 5.872807623256456e-06, | |
| "loss": 0.9447, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 5.803756387239332e-06, | |
| "loss": 0.9455, | |
| "step": 12820 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 5.734705151222207e-06, | |
| "loss": 0.9436, | |
| "step": 12840 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 5.665653915205082e-06, | |
| "loss": 0.9425, | |
| "step": 12860 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 5.596602679187958e-06, | |
| "loss": 0.9377, | |
| "step": 12880 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 5.5275514431708325e-06, | |
| "loss": 0.9464, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 5.4585002071537085e-06, | |
| "loss": 0.9432, | |
| "step": 12920 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 5.3894489711365835e-06, | |
| "loss": 0.9326, | |
| "step": 12940 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 5.320397735119459e-06, | |
| "loss": 0.9413, | |
| "step": 12960 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 5.2513464991023345e-06, | |
| "loss": 0.9353, | |
| "step": 12980 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 5.18229526308521e-06, | |
| "loss": 0.9483, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 5.116696588868941e-06, | |
| "loss": 0.934, | |
| "step": 13020 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 5.047645352851816e-06, | |
| "loss": 0.9493, | |
| "step": 13040 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 4.978594116834691e-06, | |
| "loss": 0.946, | |
| "step": 13060 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 4.909542880817567e-06, | |
| "loss": 0.9425, | |
| "step": 13080 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 4.840491644800442e-06, | |
| "loss": 0.9464, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 4.771440408783317e-06, | |
| "loss": 0.9347, | |
| "step": 13120 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 4.702389172766192e-06, | |
| "loss": 0.9393, | |
| "step": 13140 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 4.633337936749068e-06, | |
| "loss": 0.9484, | |
| "step": 13160 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 4.564286700731943e-06, | |
| "loss": 0.9284, | |
| "step": 13180 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 4.495235464714818e-06, | |
| "loss": 0.9433, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 4.426184228697694e-06, | |
| "loss": 0.9414, | |
| "step": 13220 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 4.3571329926805685e-06, | |
| "loss": 0.9464, | |
| "step": 13240 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 4.2880817566634444e-06, | |
| "loss": 0.9395, | |
| "step": 13260 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 4.21903052064632e-06, | |
| "loss": 0.9437, | |
| "step": 13280 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 4.149979284629195e-06, | |
| "loss": 0.9348, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 4.084380610412927e-06, | |
| "loss": 0.9384, | |
| "step": 13320 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 4.015329374395802e-06, | |
| "loss": 0.9448, | |
| "step": 13340 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 3.946278138378677e-06, | |
| "loss": 0.9334, | |
| "step": 13360 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 3.877226902361553e-06, | |
| "loss": 0.9362, | |
| "step": 13380 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 3.8081756663444276e-06, | |
| "loss": 0.9433, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 3.739124430327303e-06, | |
| "loss": 0.9558, | |
| "step": 13420 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 3.6700731943101786e-06, | |
| "loss": 0.9451, | |
| "step": 13440 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 3.6044745200939096e-06, | |
| "loss": 0.9362, | |
| "step": 13460 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 3.535423284076785e-06, | |
| "loss": 0.9374, | |
| "step": 13480 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 3.46637204805966e-06, | |
| "loss": 0.9326, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "eval_validation_accuracy": 0.05914255007327797, | |
| "eval_validation_loss": 0.9990234375, | |
| "eval_validation_runtime": 3709.9254, | |
| "eval_validation_samples_per_second": 0.674, | |
| "eval_validation_steps_per_second": 0.021, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 3.3973208120425356e-06, | |
| "loss": 0.9416, | |
| "step": 13520 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 3.328269576025411e-06, | |
| "loss": 0.9395, | |
| "step": 13540 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 3.259218340008286e-06, | |
| "loss": 0.9436, | |
| "step": 13560 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 3.1901671039911617e-06, | |
| "loss": 0.9387, | |
| "step": 13580 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 3.1211158679740368e-06, | |
| "loss": 0.9475, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 3.0520646319569123e-06, | |
| "loss": 0.9339, | |
| "step": 13620 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.9830133959397874e-06, | |
| "loss": 0.9387, | |
| "step": 13640 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.913962159922663e-06, | |
| "loss": 0.94, | |
| "step": 13660 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.844910923905538e-06, | |
| "loss": 0.9323, | |
| "step": 13680 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.7758596878884134e-06, | |
| "loss": 0.9402, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.7068084518712885e-06, | |
| "loss": 0.9363, | |
| "step": 13720 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.6377572158541636e-06, | |
| "loss": 0.9501, | |
| "step": 13740 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.5687059798370395e-06, | |
| "loss": 0.9503, | |
| "step": 13760 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.4996547438199146e-06, | |
| "loss": 0.938, | |
| "step": 13780 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.4306035078027896e-06, | |
| "loss": 0.9322, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.361552271785665e-06, | |
| "loss": 0.945, | |
| "step": 13820 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.2925010357685402e-06, | |
| "loss": 0.9415, | |
| "step": 13840 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.2234497997514157e-06, | |
| "loss": 0.9299, | |
| "step": 13860 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.1543985637342912e-06, | |
| "loss": 0.9413, | |
| "step": 13880 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.0853473277171663e-06, | |
| "loss": 0.9463, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.0162960917000414e-06, | |
| "loss": 0.9393, | |
| "step": 13920 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 1.947244855682917e-06, | |
| "loss": 0.9352, | |
| "step": 13940 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 1.8781936196657922e-06, | |
| "loss": 0.9319, | |
| "step": 13960 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 1.8091423836486674e-06, | |
| "loss": 0.9436, | |
| "step": 13980 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 1.7400911476315427e-06, | |
| "loss": 0.9393, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 1.671039911614418e-06, | |
| "loss": 0.9338, | |
| "step": 14020 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 1.601988675597293e-06, | |
| "loss": 0.9431, | |
| "step": 14040 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 1.5329374395801686e-06, | |
| "loss": 0.9462, | |
| "step": 14060 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 1.4638862035630439e-06, | |
| "loss": 0.9487, | |
| "step": 14080 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 1.3948349675459192e-06, | |
| "loss": 0.9335, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 1.3257837315287944e-06, | |
| "loss": 0.9268, | |
| "step": 14120 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 1.2567324955116697e-06, | |
| "loss": 0.9389, | |
| "step": 14140 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 1.187681259494545e-06, | |
| "loss": 0.9378, | |
| "step": 14160 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 1.1186300234774203e-06, | |
| "loss": 0.9277, | |
| "step": 14180 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 1.0495787874602956e-06, | |
| "loss": 0.9347, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 9.805275514431709e-07, | |
| "loss": 0.9425, | |
| "step": 14220 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 9.114763154260462e-07, | |
| "loss": 0.9396, | |
| "step": 14240 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 8.424250794089215e-07, | |
| "loss": 0.939, | |
| "step": 14260 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 7.733738433917967e-07, | |
| "loss": 0.9398, | |
| "step": 14280 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 7.04322607374672e-07, | |
| "loss": 0.9339, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 6.352713713575473e-07, | |
| "loss": 0.932, | |
| "step": 14320 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 5.662201353404226e-07, | |
| "loss": 0.9349, | |
| "step": 14340 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 4.971688993232979e-07, | |
| "loss": 0.9355, | |
| "step": 14360 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 4.281176633061732e-07, | |
| "loss": 0.9358, | |
| "step": 14380 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 3.590664272890485e-07, | |
| "loss": 0.9347, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.9001519127192375e-07, | |
| "loss": 0.9311, | |
| "step": 14420 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 2.2096395525479907e-07, | |
| "loss": 0.9445, | |
| "step": 14440 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 1.5191271923767435e-07, | |
| "loss": 0.9501, | |
| "step": 14460 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 8.286148322054965e-08, | |
| "loss": 0.9385, | |
| "step": 14480 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "step": 14484, | |
| "total_flos": 3.0144649130395304e+19, | |
| "train_loss": 0.26028791997551887, | |
| "train_runtime": 64104.6589, | |
| "train_samples_per_second": 14.46, | |
| "train_steps_per_second": 0.226 | |
| } | |
| ], | |
| "max_steps": 14484, | |
| "num_train_epochs": 3, | |
| "total_flos": 3.0144649130395304e+19, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |
