{ "run_id": "1008_qwenLfm_briage", "run_root_dir": "./results/Checkpoints", "seed": 42, "trackers": [ "jsonl", "wandb" ], "wandb_entity": "jinhuiye", "wandb_project": "InternM1", "is_debug": false, "framework": { "name": "QwenGR00T", "qwenvl": { "base_vlm": "./playground/Pretrained_models/Qwen2.5-VL-3B-Instruct", "attn_implementation": "flash_attention_2", "vl_hidden_dim": 2048 }, "dino": { "dino_backbone": "dinov2_vits14" }, "action_model": { "action_model_type": "DiT-L", "hidden_size": 1024, "add_pos_embed": true, "max_seq_len": 1024, "action_dim": 7, "state_dim": 7, "future_action_window_size": 15, "action_horizon": 16, "past_action_window_size": 0, "repeated_diffusion_steps": 8, "noise_beta_alpha": 1.5, "noise_beta_beta": 1.0, "noise_s": 0.999, "num_timestep_buckets": 1000, "num_inference_timesteps": 4, "num_target_vision_tokens": 32, "diffusion_model_cfg": { "cross_attention_dim": 2048, "dropout": 0.2, "final_dropout": true, "interleave_self_attention": true, "norm_type": "ada_norm", "num_layers": 16, "output_dim": 1024, "positional_embeddings": null }, "action_hidden_dim": 2048 } }, "datasets": { "vlm_data": { "dataset_py": "vlm_datasets", "dataformat": "llava_json", "dataset_use": "aokvqa_cauldron_llava_format,sharegpt4v_coco,sharegpt4v_knowledge,sharegpt4v_llava,sharegpt4v_sam,asv2_conversation_en,asv2_detailed_description_en,asv2_region_captioning_en,coco_internvl_longcap_en,coco_karpathy_train_567_en,coco_negative_gpt4o_en,coco_poetry_zh,coco_rem_en_zh,cocorem_exist_yorn_en,cocotextv2_en,cocotextv2_gpt4o_en,okvqa_en,refcoco_grounding_aug_en,refcoco_grounding_en,tallyqa_coco_en,toloka_grounding_aug_en,vqav2_en,vsr_en", "eval_dataset": "aokvqa_cauldron_llava_format", "data_flatten": false, "base_interval": 2, "max_pixels": 50176, "min_pixels": 784, "model_max_length": 2048, "model_type": "qwen2.5vl", "per_device_batch_size": 3 }, "vla_data": { "dataset_py": "lerobot_datasets", "data_root_dir": "playground/Datasets/OXE_LEROBOT", "data_mix": "bridge", "action_type": "delta_ee", "CoT_prompt": "Your task is {instruction}. To identify the key objects for your task. Locate their bounding boxes in [x1,y1,x2,y2] format.", "CoT_answer": "bbox", "default_image_resolution": [ 3, 224, 224 ], "per_device_batch_size": 16, "load_all_data_for_training": true, "obs": [ "image_0" ], "image_size": [ 224, 224 ] } }, "trainer": { "epochs": 100, "max_train_steps": 100000, "num_warmup_steps": 10000, "save_interval": 5000, "eval_interval": 1000, "learning_rate": { "base": 3e-05, "qwen_vl_interface": 1e-05, "action_model": 0.0001 }, "lr_scheduler_type": "cosine_with_min_lr", "scheduler_specific_kwargs": { "min_lr": 5e-07 }, "freeze_modules": true, "loss_scale": { "vla": 1.0, "vlm": 0.1 }, "repeated_diffusion_steps": 4, "max_grad_norm": 1.0, "warmup_ratio": 0.1, "weight_decay": 0.0, "logging_frequency": 10, "gradient_clipping": 1.0, "gradient_accumulation_steps": 1, "optimizer": { "name": "AdamW", "betas": [ 0.9, 0.95 ], "eps": 1e-08, "weight_decay": 1e-08 }, "is_resume": false, "resume_epoch": null, "resume_step": null, "enable_gradient_checkpointing": true, "enable_mixed_precision_training": true }, "is_resume": false, "output_dir": "./results/Checkpoints/1008_qwenLfm_briage" }