| { | |
| "best_metric": 0.2546990215778351, | |
| "best_model_checkpoint": "xlm-roberta-base-all-finetuned-toxicity-classification/checkpoint-2532", | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 7596, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.8683517640863615e-05, | |
| "loss": 0.2799, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.7367035281727228e-05, | |
| "loss": 0.2876, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.6050552922590838e-05, | |
| "loss": 0.2671, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.473407056345445e-05, | |
| "loss": 0.2639, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.3417588204318064e-05, | |
| "loss": 0.2675, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.2546990215778351, | |
| "eval_runtime": 35.0277, | |
| "eval_samples_per_second": 128.47, | |
| "eval_steps_per_second": 8.051, | |
| "step": 2532 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.2101105845181676e-05, | |
| "loss": 0.2072, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.0784623486045287e-05, | |
| "loss": 0.2035, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 9.4681411269089e-06, | |
| "loss": 0.1987, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 8.151658767772512e-06, | |
| "loss": 0.1938, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 6.835176408636125e-06, | |
| "loss": 0.189, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.29669827222824097, | |
| "eval_runtime": 35.0287, | |
| "eval_samples_per_second": 128.466, | |
| "eval_steps_per_second": 8.051, | |
| "step": 5064 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 5.518694049499738e-06, | |
| "loss": 0.1412, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 4.20221169036335e-06, | |
| "loss": 0.1467, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 2.885729331226962e-06, | |
| "loss": 0.1397, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 1.5692469720905742e-06, | |
| "loss": 0.1476, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 2.527646129541864e-07, | |
| "loss": 0.1383, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 0.4095066785812378, | |
| "eval_runtime": 35.0202, | |
| "eval_samples_per_second": 128.497, | |
| "eval_steps_per_second": 8.053, | |
| "step": 7596 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 7596, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "total_flos": 3.196799322624e+16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |