Timsty commited on
Commit
f8880d9
·
verified ·
1 Parent(s): ce452c8

Delete config.yaml

Browse files
Files changed (1) hide show
  1. config.yaml +0 -104
config.yaml DELETED
@@ -1,104 +0,0 @@
1
- run_id: 0418_QwenLatent_13tasks_actionstate_30k
2
- run_root_dir: ./runs
3
- seed: 42
4
- trackers:
5
- - jsonl
6
- - wandb
7
- wandb_entity: timsty
8
- wandb_project: vla_jepa
9
- is_debug: false
10
- framework:
11
- name: QwenLatent
12
- qwenvl:
13
- base_vlm: /mnt/data/fangyu/model/Qwen/Qwen3-VL-2B-Instruct
14
- attn_implementation: flash_attention_2
15
- vl_hidden_dim: 2048
16
- num_data_tokens: 32
17
- action_model:
18
- ckpt_path: /mnt/data/fangyu/code/reward_new/runs/0418_Action_13tasks_actionstate_fixchunk15/final_model/pytorch_model.pt
19
- action_size: 37
20
- state_size: 74
21
- use_state: ${datasets.vla_data.state_use_action_chunk}
22
- hidden_size: 1024
23
- intermediate_size: 3072
24
- dataset_vocab_size: 256
25
- num_data_tokens: 32
26
- num_t_samples: 4
27
- min_action_len: 5
28
- num_encoder_layers: 28
29
- num_decoder_layers: 28
30
- num_attention_heads: 16
31
- num_key_value_heads: 8
32
- head_dim: 128
33
- max_position_embeddings: 2048
34
- max_action_chunk_size: 50
35
- rms_norm_eps: 1.0e-06
36
- attention_dropout: 0.0
37
- use_vae_reparameterization: false
38
- use_ema: false
39
- chunk_size: ${datasets.vla_data.chunk_size}
40
- loss_mode: full
41
- qwen3_pretrained_name_or_path: /mnt/data/fangyu/model/Qwen/Qwen3-0.6B
42
- datasets:
43
- vla_data:
44
- dataset_py: lerobot_datasets
45
- data_root_dir: /mnt/data/fangyu/dataset/IPEC-COMMUNITY
46
- data_mix: cross_embodiedment_13tasks
47
- CoT_prompt: 'Task: {instruction}. What are the next 15 actions to take?'
48
- default_image_resolution:
49
- - 3
50
- - 224
51
- - 224
52
- per_device_batch_size: 32
53
- load_all_data_for_training: true
54
- obs:
55
- - image_0
56
- image_size:
57
- - 224
58
- - 224
59
- video_backend: torchcodec
60
- load_video: true
61
- chunk_size: 15
62
- state_use_action_chunk: true
63
- num_history_steps: 0
64
- include_state: ${datasets.vla_data.state_use_action_chunk}
65
- trainer:
66
- epochs: 100
67
- max_train_steps: 50000
68
- num_warmup_steps: 5000
69
- num_stable_steps: 0
70
- mode: decay_aux_loss
71
- loss_weights_decay_steps: 5000
72
- save_interval: 5000
73
- eval_interval: 50
74
- max_checkpoints_to_keep: 20
75
- learning_rate:
76
- base: 2.5e-05
77
- qwen_vl_interface: 2.5e-05
78
- action_model: 2.5e-05
79
- lr_scheduler_type: warmup_stable_cosine
80
- scheduler_specific_kwargs:
81
- min_lr_ratio: 0.001
82
- freeze_modules: ''
83
- loss_scale:
84
- align_loss: 1.0
85
- recon_loss: 1.0
86
- predict_loss: 1.0
87
- warmup_ratio: 0.1
88
- weight_decay: 0.0
89
- logging_frequency: 10
90
- gradient_clipping: 5.0
91
- gradient_accumulation_steps: 1
92
- optimizer:
93
- name: AdamW
94
- betas:
95
- - 0.9
96
- - 0.95
97
- eps: 1.0e-08
98
- weight_decay: 1.0e-08
99
- is_resume: false
100
- resume_epoch: null
101
- resume_step: null
102
- enable_gradient_checkpointing: true
103
- enable_mixed_precision_training: true
104
- output_dir: ./runs/0418_QwenLatent_13tasks_actionstate_30k