{ "batch_size": 1, "accum_iter": 8, "epochs": 100, "warmup_epochs": 0.001, "lr": 0.0002, "min_lr": 0.0002, "wd": 0.1, "clip_grad": 4.0, "init_from": "/mnt/petrelfs/gaopeng/synbol/mgpt-dev-main-7B-768/lumina_mgpt/output/7B_2e-4_1_10240_768_64A100_final_stage/epoch0-iter20999", "data_config": "/mnt/petrelfs/gaopeng/synbol/mgpt-dev-main-7B-768/lumina_mgpt/configs/data/sample.yaml", "cache_ann_on_disk": true, "length_clustering": true, "num_workers": 8, "pin_mem": true, "seed": 0, "output_dir": "output/7B_2e-4_1_10240_768_64A100_final_stage", "save_interval": 1, "save_iteration_interval": 1000, "only_save_trainable": false, "ckpt_max_keep": 2, "auto_resume": true, "resume_path": "output/7B_2e-4_1_10240_768_64A100_final_stage/epoch0-iter20999", "model_parallel_size": 1, "data_parallel": "fsdp", "precision": "bf16", "grad_precision": "fp32", "checkpointing": true, "max_seq_len": 10240, "mask_image_logits": false, "dropout": 0.05, "z_loss_weight": 1e-05, "model_size": "7B", "world_size": 64, "rank": 0, "gpu": 0, "local_rank": 0, "dist_url": "env://", "distributed": true, "dist_backend": "nccl" }