{ "best_metric": 1.0803709030151367, "best_model_checkpoint": "./outputs/llava-mistral/RLAIF-V_Coocur-q0_50/checkpoint-100", "epoch": 0.6956521739130435, "eval_steps": 50, "global_step": 100, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.034782608695652174, "grad_norm": 28.788983552687714, "learning_rate": 5e-07, "loss": 1.6631, "step": 5 }, { "epoch": 0.06956521739130435, "grad_norm": 14.801312030106265, "learning_rate": 1e-06, "loss": 1.5311, "step": 10 }, { "epoch": 0.10434782608695652, "grad_norm": 8.042597892857103, "learning_rate": 9.99648681635985e-07, "loss": 1.2909, "step": 15 }, { "epoch": 0.1391304347826087, "grad_norm": 5.465265266311333, "learning_rate": 9.985952202423114e-07, "loss": 1.2361, "step": 20 }, { "epoch": 0.17391304347826086, "grad_norm": 5.471328096600488, "learning_rate": 9.96841096220313e-07, "loss": 1.2008, "step": 25 }, { "epoch": 0.20869565217391303, "grad_norm": 5.088536561695133, "learning_rate": 9.943887745939163e-07, "loss": 1.184, "step": 30 }, { "epoch": 0.24347826086956523, "grad_norm": 4.979056344341446, "learning_rate": 9.912417015456088e-07, "loss": 1.1743, "step": 35 }, { "epoch": 0.2782608695652174, "grad_norm": 5.098810591402351, "learning_rate": 9.874042995736093e-07, "loss": 1.1533, "step": 40 }, { "epoch": 0.3130434782608696, "grad_norm": 4.9260826588669575, "learning_rate": 9.828819612770495e-07, "loss": 1.15, "step": 45 }, { "epoch": 0.34782608695652173, "grad_norm": 5.009097422362774, "learning_rate": 9.77681041777897e-07, "loss": 1.1191, "step": 50 }, { "epoch": 0.34782608695652173, "eval_loss": 1.126644253730774, "eval_runtime": 71.6515, "eval_samples_per_second": 57.026, "eval_steps_per_second": 0.893, "step": 50 }, { "epoch": 0.3826086956521739, "grad_norm": 4.710201613179453, "learning_rate": 9.718088497902707e-07, "loss": 1.1278, "step": 55 }, { "epoch": 0.41739130434782606, "grad_norm": 4.656292887228332, "learning_rate": 9.652736373497e-07, "loss": 1.1259, "step": 60 }, { "epoch": 0.45217391304347826, "grad_norm": 4.89840217135744, "learning_rate": 9.580845882167572e-07, "loss": 1.1196, "step": 65 }, { "epoch": 0.48695652173913045, "grad_norm": 5.131971560708056, "learning_rate": 9.502518049713631e-07, "loss": 1.0982, "step": 70 }, { "epoch": 0.5217391304347826, "grad_norm": 5.025459861546626, "learning_rate": 9.417862948158997e-07, "loss": 1.1104, "step": 75 }, { "epoch": 0.5565217391304348, "grad_norm": 5.083697590841828, "learning_rate": 9.326999541070803e-07, "loss": 1.0856, "step": 80 }, { "epoch": 0.591304347826087, "grad_norm": 5.065321153827992, "learning_rate": 9.23005551638316e-07, "loss": 1.1057, "step": 85 }, { "epoch": 0.6260869565217392, "grad_norm": 4.793827429409211, "learning_rate": 9.127167106960681e-07, "loss": 1.0966, "step": 90 }, { "epoch": 0.6608695652173913, "grad_norm": 4.988686751743354, "learning_rate": 9.018478899154066e-07, "loss": 1.0984, "step": 95 }, { "epoch": 0.6956521739130435, "grad_norm": 5.137325195769105, "learning_rate": 8.904143629616732e-07, "loss": 1.0827, "step": 100 }, { "epoch": 0.6956521739130435, "eval_loss": 1.0803709030151367, "eval_runtime": 71.4447, "eval_samples_per_second": 57.191, "eval_steps_per_second": 0.896, "step": 100 } ], "logging_steps": 5, "max_steps": 429, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 589411381149696.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }