Lumina-mGPT-2.0 / args.json
XiN0919's picture
Upload folder using huggingface_hub
4947510 verified
{
"batch_size": 1,
"accum_iter": 8,
"epochs": 100,
"warmup_epochs": 0.001,
"lr": 0.0002,
"min_lr": 0.0002,
"wd": 0.1,
"clip_grad": 4.0,
"init_from": "/mnt/petrelfs/gaopeng/synbol/mgpt-dev-main-7B-768/lumina_mgpt/output/7B_2e-4_1_10240_768_64A100_final_stage/epoch0-iter20999",
"data_config": "/mnt/petrelfs/gaopeng/synbol/mgpt-dev-main-7B-768/lumina_mgpt/configs/data/sample.yaml",
"cache_ann_on_disk": true,
"length_clustering": true,
"num_workers": 8,
"pin_mem": true,
"seed": 0,
"output_dir": "output/7B_2e-4_1_10240_768_64A100_final_stage",
"save_interval": 1,
"save_iteration_interval": 1000,
"only_save_trainable": false,
"ckpt_max_keep": 2,
"auto_resume": true,
"resume_path": "output/7B_2e-4_1_10240_768_64A100_final_stage/epoch0-iter20999",
"model_parallel_size": 1,
"data_parallel": "fsdp",
"precision": "bf16",
"grad_precision": "fp32",
"checkpointing": true,
"max_seq_len": 10240,
"mask_image_logits": false,
"dropout": 0.05,
"z_loss_weight": 1e-05,
"model_size": "7B",
"world_size": 64,
"rank": 0,
"gpu": 0,
"local_rank": 0,
"dist_url": "env://",
"distributed": true,
"dist_backend": "nccl"
}