{ "best_metric": 1.0978167057037354, "best_model_checkpoint": "./outputs/llava-mistral/RLAIF-V_Coocur-q0_25/checkpoint-100", "epoch": 1.3888888888888888, "eval_steps": 50, "global_step": 100, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.06944444444444445, "grad_norm": 28.445383103621488, "learning_rate": 5e-07, "loss": 1.692, "step": 5 }, { "epoch": 0.1388888888888889, "grad_norm": 15.214710637322117, "learning_rate": 1e-06, "loss": 1.5428, "step": 10 }, { "epoch": 0.20833333333333334, "grad_norm": 7.725515360957365, "learning_rate": 9.985471028179154e-07, "loss": 1.3232, "step": 15 }, { "epoch": 0.2777777777777778, "grad_norm": 5.771045580757361, "learning_rate": 9.94196854912548e-07, "loss": 1.2575, "step": 20 }, { "epoch": 0.3472222222222222, "grad_norm": 5.8072030206153045, "learning_rate": 9.869745381355905e-07, "loss": 1.2163, "step": 25 }, { "epoch": 0.4166666666666667, "grad_norm": 5.2740216973392755, "learning_rate": 9.769221256218162e-07, "loss": 1.1937, "step": 30 }, { "epoch": 0.4861111111111111, "grad_norm": 5.030592874663904, "learning_rate": 9.64098037858483e-07, "loss": 1.1473, "step": 35 }, { "epoch": 0.5555555555555556, "grad_norm": 4.82797980028618, "learning_rate": 9.485768031694871e-07, "loss": 1.1529, "step": 40 }, { "epoch": 0.625, "grad_norm": 5.01413477732448, "learning_rate": 9.304486245873971e-07, "loss": 1.1598, "step": 45 }, { "epoch": 0.6944444444444444, "grad_norm": 5.005624161351076, "learning_rate": 9.098188556305262e-07, "loss": 1.1343, "step": 50 }, { "epoch": 0.6944444444444444, "eval_loss": 1.1338481903076172, "eval_runtime": 35.4083, "eval_samples_per_second": 57.698, "eval_steps_per_second": 0.904, "step": 50 }, { "epoch": 0.7638888888888888, "grad_norm": 5.121349492890328, "learning_rate": 8.868073880316123e-07, "loss": 1.1465, "step": 55 }, { "epoch": 0.8333333333333334, "grad_norm": 4.967730285097845, "learning_rate": 8.615479549763755e-07, "loss": 1.1205, "step": 60 }, { "epoch": 0.9027777777777778, "grad_norm": 4.888543325436437, "learning_rate": 8.341873539012443e-07, "loss": 1.1101, "step": 65 }, { "epoch": 0.9722222222222222, "grad_norm": 5.136971847854665, "learning_rate": 8.048845933670271e-07, "loss": 1.1137, "step": 70 }, { "epoch": 1.0416666666666667, "grad_norm": 5.65186903458913, "learning_rate": 7.738099689665539e-07, "loss": 1.0344, "step": 75 }, { "epoch": 1.1111111111111112, "grad_norm": 5.436495818544836, "learning_rate": 7.41144073636728e-07, "loss": 0.9903, "step": 80 }, { "epoch": 1.1805555555555556, "grad_norm": 5.1217292921898485, "learning_rate": 7.070767481266492e-07, "loss": 0.976, "step": 85 }, { "epoch": 1.25, "grad_norm": 5.550754536917889, "learning_rate": 6.718059777212565e-07, "loss": 0.9731, "step": 90 }, { "epoch": 1.3194444444444444, "grad_norm": 5.3291194115834095, "learning_rate": 6.355367416322778e-07, "loss": 0.9687, "step": 95 }, { "epoch": 1.3888888888888888, "grad_norm": 5.187329472162837, "learning_rate": 5.984798217433531e-07, "loss": 0.9522, "step": 100 }, { "epoch": 1.3888888888888888, "eval_loss": 1.0978167057037354, "eval_runtime": 35.0302, "eval_samples_per_second": 58.321, "eval_steps_per_second": 0.913, "step": 100 } ], "logging_steps": 5, "max_steps": 216, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 589411381149696.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }