{ "act_dim": 3, "action_tanh": true, "activation_function": "relu", "adv_act_dim": 3, "architectures": [ "MultipartARDT" ], "attn_pdrop": 0.1, "bos_token_id": 50256, "context_size": 20, "embd_pdrop": 0.1, "eos_token_id": 50256, "flag": 0, "hidden_size": 128, "initializer_range": 0.02, "lambda1": 1.0, "lambda2": 1.0, "layer_norm_epsilon": 1e-05, "log_interval_steps": 100, "max_ep_len": 999, "max_ep_return": 3072.5702787999967, "max_obs_len": 999, "max_obs_return": 3072.5702787999967, "min_ep_return": 0.59533596, "min_obs_return": 0.59533596, "model_type": "decision_transformer", "n_head": 1, "n_inner": null, "n_layer": 3, "n_positions": 1024, "pr_act_dim": 3, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "returns_scale": 1000, "rtg_shift": 0, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "state_dim": 11, "state_mean": [ 1.2484109448283487, 0.02541937343443989, -0.417574769829802, -0.1535401766461714, 0.12844594577746832, 1.7648832092997122, -0.14629294202424897, 0.005822005615959526, -0.3319174318352961, -0.14299996508458745, -0.5025578820838597 ], "state_std": [ 0.14913521175340505, 0.08197128929621504, 0.2847277032366293, 0.2506871380229607, 0.5970928975497837, 0.903718761191001, 1.2132053487487453, 0.9225378063227366, 1.897097236674443, 2.661700770003696, 6.165093175492324 ], "torch_dtype": "float32", "total_train_steps": 10000, "transformers_version": "4.31.0", "use_cache": true, "vocab_size": 1, "warmup_steps": 1000 }