text stringlengths 0 4.99k |
|---|
inputs = layers.Input(shape=(num_states,)) |
out = layers.Dense(256, activation=\"relu\")(inputs) |
out = layers.Dense(256, activation=\"relu\")(out) |
outputs = layers.Dense(1, activation=\"tanh\", kernel_initializer=last_init)(out) |
# Our upper bound is 2.0 for Pendulum. |
outputs = outputs * upper_bound |
model = tf.keras.Model(inputs, outputs) |
return model |
def get_critic(): |
# State as input |
state_input = layers.Input(shape=(num_states)) |
state_out = layers.Dense(16, activation=\"relu\")(state_input) |
state_out = layers.Dense(32, activation=\"relu\")(state_out) |
# Action as input |
action_input = layers.Input(shape=(num_actions)) |
action_out = layers.Dense(32, activation=\"relu\")(action_input) |
# Both are passed through seperate layer before concatenating |
concat = layers.Concatenate()([state_out, action_out]) |
out = layers.Dense(256, activation=\"relu\")(concat) |
out = layers.Dense(256, activation=\"relu\")(out) |
outputs = layers.Dense(1)(out) |
# Outputs single value for give state-action |
model = tf.keras.Model([state_input, action_input], outputs) |
return model |
policy() returns an action sampled from our Actor network plus some noise for exploration. |
def policy(state, noise_object): |
sampled_actions = tf.squeeze(actor_model(state)) |
noise = noise_object() |
# Adding noise to action |
sampled_actions = sampled_actions.numpy() + noise |
# We make sure action is within bounds |
legal_action = np.clip(sampled_actions, lower_bound, upper_bound) |
return [np.squeeze(legal_action)] |
Training hyperparameters |
std_dev = 0.2 |
ou_noise = OUActionNoise(mean=np.zeros(1), std_deviation=float(std_dev) * np.ones(1)) |
actor_model = get_actor() |
critic_model = get_critic() |
target_actor = get_actor() |
target_critic = get_critic() |
# Making the weights equal initially |
target_actor.set_weights(actor_model.get_weights()) |
target_critic.set_weights(critic_model.get_weights()) |
# Learning rate for actor-critic models |
critic_lr = 0.002 |
actor_lr = 0.001 |
critic_optimizer = tf.keras.optimizers.Adam(critic_lr) |
actor_optimizer = tf.keras.optimizers.Adam(actor_lr) |
total_episodes = 100 |
# Discount factor for future rewards |
gamma = 0.99 |
# Used to update target networks |
tau = 0.005 |
buffer = Buffer(50000, 64) |
Now we implement our main training loop, and iterate over episodes. We sample actions using policy() and train with learn() at each time step, along with updating the Target networks at a rate tau. |
# To store reward history of each episode |
ep_reward_list = [] |
# To store average reward history of last few episodes |
avg_reward_list = [] |
# Takes about 4 min to train |
for ep in range(total_episodes): |
prev_state = env.reset() |
episodic_reward = 0 |
while True: |
# Uncomment this to see the Actor in action |
# But not in a python notebook. |
# env.render() |
tf_prev_state = tf.expand_dims(tf.convert_to_tensor(prev_state), 0) |
action = policy(tf_prev_state, ou_noise) |
# Recieve state and reward from environment. |
state, reward, done, info = env.step(action) |
buffer.record((prev_state, action, reward, state)) |
episodic_reward += reward |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.