data_root_dir: /hai/scratch/belkhale/datasets hf_token: .hf_token image_aug: false is_resume: true pretrained_checkpoint: runs/prism-qwen25-dinosiglip-224px+0_5b+mx-libero-90+n1+b32+x7--vq_extra_tokenizer/checkpoints/step-065000-epoch-29-loss=0.4407.pt resume_epoch: 29 resume_step: 65000 run_id: prism-qwen25-dinosiglip-224px+0_5b+mx-libero-90+n1+b32+x7--vq_extra_tokenizer run_id_note: vq_extra_tokenizer run_root_dir: runs save_interval: 2500 seed: 7 trackers: - jsonl - wandb vla: action_tokenizer: libero_vq_extra_action_tokenizer base_vlm: prism-qwen25-extra-dinosiglip-224px+0_5b data_mix: libero_90 enable_gradient_checkpointing: true enable_mixed_precision_training: true epochs: 1000 expected_world_size: 8 freeze_llm_backbone: false freeze_vision_backbone: false global_batch_size: 256 learning_rate: 2.0e-05 lr_scheduler_type: constant max_grad_norm: 1.0 max_steps: null per_device_batch_size: 32 reduce_in_full_precision: true save_every_n_steps: 25000 shuffle_buffer_size: 256000 train_strategy: fsdp-full-shard type: prism-qwen25-dinosiglip-224px+0_5b+mx-libero-90 unfreeze_last_llm_layer: false vla_id: prism-qwen25-dinosiglip-224px+0_5b+mx-libero-90 warmup_ratio: 0.0 weight_decay: 0.0 wandb_entity: null wandb_project: prismatic