hpt-small / config.yaml
liruiw's picture
Upload 3 files
2ad14f6 verified
seed: 0
output_dir: output/04_04_2024_00_18_28_1061204
domains: austin_sailor_dataset_converted_externally_to_rlds, stanford_hydra_dataset_converted_externally_to_rlds,
austin_buds_dataset_converted_externally_to_rlds, austin_sirius_dataset_converted_externally_to_rlds,
berkeley_mvp_converted_externally_to_rlds, berkeley_rpt_converted_externally_to_rlds,
iamlab_cmu_pickup_insert_converted_externally_to_rlds, utaustin_mutex, imperialcollege_sawyer_wrist_cam,
stanford_mask_vit_converted_externally_to_rlds, language_table, kuka, bc_z, robo_net,
dlr_sara_pour_converted_externally_to_rlds, stanford_robocook_converted_externally_to_rlds,
cmu_play_fusion, bridge, furniture_bench_dataset_converted_externally_to_rlds, ucsd_pick_and_place_dataset_converted_externally_to_rlds,
usc_cloth_sim_converted_externally_to_rlds, stanford_kuka_multimodal_dataset_converted_externally_to_rlds,
roboturk, kaist_nonprehensile_converted_externally_to_rlds, asu_table_top_converted_externally_to_rlds,
utokyo_xarm_pick_and_place_converted_externally_to_rlds, berkeley_cable_routing,
droid
log_dir: output/04_04_2024_00_18_28_1061204
debug_distributed: false
wb_tag: default
wb_cont_run: 37fe2kep
log_interval: 10
script_name: run_resnet_30dataset_traj1000_embed_128_batch1024_800k_cons
save_wb_checkpoint: true
slurm_job_id: '25682608'
effective_total_epochs: 100
effective_batch_size: 256
epoch_size: 10
total_num_traj: 0
total_num_sample: 0
rank: 0
gpu: 0
task_per_gpu: 1
world_size: 32
debug_submitit: false
ngpus: 8
nodes: 4
timeout: 4320
job_dir: logs/
partition: learnlab
use_volta32: true
comment: ''
resume: logs/
dist_url: file:///checkpoint/xinleic/experiments/b96f1c1c8a1e49f691901ccebe7e3f1f_init
dist_on_itp: false
local_rank: 1
distributed: true
dist_backend: nccl
dset_w_temperature: 2.0
dataset_shuffle: true
dataset_groups: ''
nodelist: learnlab,learnfair,scavenge
fsdp: false
dataset:
_target_: hpt_pretrain.dataset.traj_dataset.TrajDataset
horizon: 5
val_ratio: 0.1
pad_after: 0
precompute_feat: true
image_encoder: resnet
episode_cnt: 1000
step_cnt: 10000000
data_augmentation: false
use_disk: true
pad_before: 0
data_ratio: 1
action_horizon: 8
observation_horizon: 4
dataset_postfix: _traj1000
dataset_encoder_postfix: _resnet
use_multiview: false
normalize_state: true
use_heldout_dataset: true
heldout_dataset: false
regenerate: false
continue_generate: false
network:
_target_: hpt_pretrain.models.policy.Policy
embed_dim: 128
num_blocks: 16
num_heads: 8
use_modality_embedding: true
use_domain_embedding: false
token_postprocessing: mean
weight_init_style: pytorch
drop_path: 0.1
mae_loss_scale: 0.0
masked_autoencoding: false
stem:
modalities:
- image
- state
modality_embed_dim: 128
normalize_state: ${dataset.normalize_state}
state_embedding_dim: 1
image_encoder: ${dataset.image_encoder}
crossattn_dim_head: 64
crossattn_heads: 8
crossattn_modality_dropout: 0.1
observation_horizon: ${dataset.observation_horizon}
random_horizon_masking: true
add_pos_embedding_to_state: false
num_blocks: 1
crossattn_latent:
image: 16
state: 16
image:
_target_: hpt_pretrain.models.policy_stem.MLP
input_dim: 512
output_dim: 128
widths:
- 128
num_of_copy: 1
state:
_target_: hpt_pretrain.models.policy_stem.MLP
input_dim: 7
output_dim: 128
widths:
- 128
head:
_target_: hpt_pretrain.models.policy_head.MLP
input_dim: 128
tanh_end: true
output_dim: 56
dropout: true
widths:
- 256
- 128
dataloader:
batch_size: 32
num_workers: 1
pin_memory: false
persistent_workers: false
drop_last: true
val_dataloader:
num_workers: 1
pin_memory: false
persistent_workers: false
ddp_dataloader:
num_workers: 8
pin_memory: false
persistent_workers: false
drop_last: false
prefetch_factor: 2
ddp_val_dataloader:
num_workers: 8
pin_memory: false
persistent_workers: false
drop_last: false
prefetch_factor: 2
optimizer:
_target_: torch.optim.AdamW
lr: 0.0008
eps: 1.0e-06
weight_decay: 0.05
optimizer_misc:
nontrunk_lr_scale: 0.5
warmup_lr:
lr: 1.0e-10
step: 1000
train:
total_epochs: 3000
total_iters: 800000
epoch_iters: 1000
validation_iters: 100
use_accumulation: false
pretrained_dir: ''
max_validation_size: 10
accumulate_batch_step: 1
lr_scheduler:
_target_: torch.optim.lr_scheduler.ConstantLR
factor: 1