default_settings: null behaviors: Pyramids: trainer_type: ppo hyperparameters: batch_size: 128 buffer_size: 2048 learning_rate: 0.0003 beta: 0.01 epsilon: 0.2 lambd: 0.95 num_epoch: 3 learning_rate_schedule: linear beta_schedule: linear epsilon_schedule: linear network_settings: normalize: false hidden_units: 512 num_layers: 2 vis_encode_type: simple memory: null goal_conditioning_type: hyper deterministic: false reward_signals: extrinsic: gamma: 0.99 strength: 1.0 network_settings: normalize: false hidden_units: 128 num_layers: 2 vis_encode_type: simple memory: null goal_conditioning_type: hyper deterministic: false rnd: gamma: 0.99 strength: 0.02 network_settings: normalize: false hidden_units: 64 num_layers: 3 vis_encode_type: simple memory: null goal_conditioning_type: hyper deterministic: false learning_rate: 0.0001 encoding_size: null init_path: null keep_checkpoints: 5 checkpoint_interval: 200000 max_steps: 1000000 time_horizon: 128 summary_freq: 30000 threaded: false self_play: null behavioral_cloning: null env_settings: env_path: ./training-envs-executables/linux/Pyramids/Pyramids env_args: null base_port: 5005 num_envs: 1 num_areas: 1 seed: -1 max_lifetime_restarts: 10 restarts_rate_limit_n: 1 restarts_rate_limit_period_s: 60 engine_settings: width: 84 height: 84 quality_level: 5 time_scale: 20 target_frame_rate: -1 capture_frame_rate: 60 no_graphics: true environment_parameters: null checkpoint_settings: run_id: Pyramids Training initialize_from: null load_model: false resume: false force: false train_model: false inference: false results_dir: results torch_settings: device: null debug: false