File size: 23,229 Bytes
eb3678a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
# Cross The Road Env Tutorial

Learn how to use Godot RL Agents by building a grid-like cross-the-road mini-game that will be trained using Deep Reinforcement Learning!

## **Introduction:**

Welcome to this tutorial, where you will train a robot to reach the goal while navigating around trees and avoiding cars.

At the end, you will have a trained agent that can complete the game like in the video:

https://github.com/user-attachments/assets/17fb94e6-771e-42ea-a530-cea028545a67

## Objectives:

- Learn how to use Godot RL Agents by training an agent to complete a grid-like mini-game environment.

## Prerequisites and requirements:

- Recommended: Complete [Godot RL Agents](https://huggingface.co/learn/deep-rl-course/unitbonus3/godotrl) before starting this tutorial,
- Some familiarity with Godot is also recommended,
- Godot 4.3 with .NET support (tested on Beta 3, should work with newer versions),
- Godot RL Agents (you can use `pip install godot-rl` in the venv/conda env)

# The environment

![the_environment](https://github.com/user-attachments/assets/b47aec6c-0112-4ec1-898c-209ec2652cf6)

- In the environment, the robot needs to navigate around trees and cars in order to reach the goal tile successfully within the time limit.
- The positions of goal, trees and car starting position/orientation are randomized within their own grid row. The env layout can be further customized in the code (we’ll cover this in the later sections), and you can e.g. add more cars and trees to make the game more challenging.
- As it is a grid-like env, all movements are done by 1 grid coordinate at a time, and the episode lengths are relatively short.

# Getting started:

To get started, download the project by cloning [this](https://github.com/edbeeching/godot_rl_agents_examples/tree/main) repository.

We will work on the starter project, focusing on:

- Implementing the code for the `Robot` and `AIController` nodes,
- We will briefly go over how to adjust the map size,
- Finally we will train the agent and export an .onnx file of the trained agent, so that we can run inference directly from Godot.

### Open the starter project in Godot

Open Godot, click “Import” and navigate to the `Starter` folder `godot_rl_agents_examples/examples/CrossTheRoad

/Starter/`.



On first import, you may get an error message such as: 



> Unable to load addon script from path: 'res://addons/godot_rl_agents/godot_rl_agents.gd'. This might be due to a code error in that script.

Disabling the addon at 'res://addons/godot_rl_agents/plugin.cfg' to prevent further errors.

    

Just re-open the project in the editor, the error should not appear again. 



### Open the `robot.tscn` scene



> [!TIP]

> You can search for “robot” in the FileSystem search.



<img width="544" alt="robot_scene" src="https://github.com/user-attachments/assets/315fe2b9-c107-4ff0-84a9-4c5fc9352838">



This scene contains a few nodes:

- `Robot` is the main node that controls the robot

- `AIController3D` handles observations, rewards, and setting actions,

- `robot` contains the visual geometry and animation.



### Open the `Robot.gd` script for editing:



> [!TIP]

> To open the script, click on the scroll next to the `Robot` node.



You will find there are commented method names in the script that need to be implemented. Most of the provided code will contain comments explaining the implementation, but we will also provide an additional overview of the methods.



**Replace `_ready()`, `_physics_process(delta)`, `_process_movement(_delta)` and `reward_approaching_goal()` with the following implementation:**



```gdscript

func _ready():
	reset()


func _physics_process(delta):
	# Set to true by the sync node when reset is requested from Python (starting training, evaluation, etc.)

	if _ai_controller.needs_reset:

		game_over()

	_process_movement(delta)


func _process_movement(_delta):

	for car in car_manager.cars:
		if get_grid_position() == map.get_grid_position(car.global_position):

			# If a car has moved to the current player position, end episode

			game_over(0)

			print_game_status("Failed, hit car while standing")


	if requested_movement:

		# Move the robot to the requested position

		global_position += (requested_movement * movement_step)

		# Update the visual rotation of the robot to look toward the direction of last requested movement

		visual_robot.global_rotation_degrees.y = rad_to_deg(atan2(-requested_movement.x, -requested_movement.z))

		

		var grid_position: Vector3i = get_grid_position()

		var tile: Tile = map.get_tile(grid_position)

		

		if not tile:

			# Push the robot back if there's no tile to move to (out of map boundary)

			global_position -= (requested_movement * movement_step)

		elif tile.id == tile.TileNames.tree:

			# Push the robot back if it has moved to a tree tile

			global_position -= (requested_movement * movement_step)

		elif tile.id == tile.TileNames.goal:

			# If a goal tile is reached, end episode with reward +1

			game_over(1)

			print_game_status("Success, reached goal")

		else:

			for car in car_manager.cars:

				if get_grid_position() == map.get_grid_position(car.global_position):

					# If the robot moved to a car's current position, end episode

					game_over(0)

					print_game_status("Failed, hit car while walking")


		# After processing the move, zero the movement for the next step

		# (only in case of human control)

		if _ai_controller.control_mode == AIController3D.ControlModes.HUMAN:

			requested_movement = Vector3.ZERO 

			

	reward_approaching_goal()


## Adds a positive reward if the robot approaches the goal
func reward_approaching_goal():
	var grid_pos: Vector3i = get_grid_position()

	var dist_to_goal = grid_pos.distance_to(map.goal_position)

	if last_dist_to_goal == null: last_dist_to_goal = dist_to_goal

	

	if dist_to_goal < last_dist_to_goal:

		_ai_controller.reward += (last_dist_to_goal - dist_to_goal) / 10.0

		last_dist_to_goal = dist_to_goal

```


A brief overview of the methods:

- In `_ready()` we reset the game on start.
- In `_process_movement()` we move the Robot, handle collision with tree tiles, goal (in this case we add a positive reward and end the episode), and cars (if a car is hit, we add 0 to the reward and end the episode).
- In `reward_approaching_goal()` we give a positive reward when the robot approaches the goal, based on best/smallest distance to the goal, which is reset for each episode/game.

**Replace `get_grid_position()`, `game_over()`, `reset()`, and `print_game_status()` with the following implementation:**



```gdscript

func get_grid_position() -> Vector3i:

	return map.get_grid_position(global_position)



func game_over(reward = 0.0):

	_ai_controller.done = true

	_ai_controller.reward += reward

	_ai_controller.reset()

	reset()



func reset():

	last_dist_to_goal = null

	# Order of resetting is important:

	# We reset the map first, which sets a new player start position

	# and the road segments (needed to know where to spawn the cars)

	map.reset()

	# after that, we can set the player position

	global_position = Vector3(map.player_start_position) + Vector3.UP * 1.5

	# and also reset or create (on first start) the cars

	car_manager.reset()



func print_game_status(message: String):

	if print_game_status_enabled:

		print(message)

```



A brief overview of the methods:



- `get_grid_position()` tells us where the car is on the grid.

- `game_over()` takes care of resetting the player, but also ending the episode for the RL agent and resetting the AIController, and the environment (e.g. generating a new map for the next episode) as well.

- `print_game_status()` prints out game status messages (e.g. game successfully completed) in the console when this flag is enabled on the Player node in the inspector. This is disabled by default.



That’s it for the Player node. Press `CTRL + S` to save the changes, and let’s implement the AIController code.



### Open the `robot_ai_controller.gd` script for editing:



> [!TIP]

> To open the script, click on the scroll next to the `AIController3D` node.



**Replace `get_obs()` with the following implementation:**

```gdscript

func get_obs() -> Dictionary:

	var observations := Array()

	var player_pos = player.get_grid_position()



	# Determines how many grid cells the AI can observe

	# e.g. front refers to "up" looking from above,

	# and does not rotate with the robot, same with others.

	# Set to observe all cells from two rows in front of player,

	# 0 behind

	var visible_rows_in_front_of_player: int = 2

	var visible_rows_behind_the_player: int = 0



	# Set to observe the entire width of the grid on both sides

	# so it will always see all cells up to 2 rows in front.

	var visible_columns_left_of_player: int = player.map.grid_size_x

	var visible_columns_right_of_player: int = player.map.grid_size_x



	# For tiles near player we provide [direction, id] (e.g: [-1, 0])

	# direction is -1 or 1 for cars, 0 for static tiles

	# if a car is in a tile, we override the id of tile underneath



	# Car ID is the ID of the last tile + 1

	var car_id = Tile.TileNames.size()

	

	# If there is no tile placed at the grid coord, we use -1 as id

	var no_tile_id: int = -1



	# Note: We don't need to include the player position directly in obs

	# as we are always including data for cells "around" the player,

	# so the player location relative to those cells is implicitly included



	for z in range(

			player_pos.z - visible_rows_in_front_of_player,

			player_pos.z + visible_rows_behind_the_player + 1

		):

		for x in range(

				player_pos.x - visible_columns_left_of_player,

				player_pos.x + visible_columns_right_of_player + 1

			):

			var grid_pos := Vector3i(x, 0, z)

			var tile: Tile = player.map.get_tile(grid_pos)



			if not tile:

				observations.append_array([0, no_tile_id])

			else:

				var is_car: bool

				for car in player.car_manager.cars:

					if grid_pos == player.map.get_grid_position(car.global_position):

						is_car = true

						observations.append(car.current_direction)

				if is_car:

					observations.append(car_id)

				else:

					observations.append_array([0, tile.id])



	return {"obs": observations}

```

In this method, we provide observations to the RL agent about the grid tiles (or cars) that are near the player. As info from grid cells near the player is provided, we don’t need to include the player position in the grid directly in the obs.

With the implementation above, the agent can “see" up to two rows in front of it (toward the goal), zero rows behind, and all of the cells left/right from the player are visible. You can adjust this as needed if you change the map size, or to experiment.

For each cell, we provide a `tile id` (e.g. 0, 1, 2, 3) so that the agent can differentiate between different tiles. If there’s a car above a specific tile, we provide the `car id` instead of the tile id, so that the agent knows there’s a car to avoid there. As cars can move and change direction, we also provide the direction (-1 or 1 for cars, 0 for tiles). We also provide `-1` as id if there’s nothing at that coordinate (out of grid boundaries).

**Replace `get_reward()`, `_process()`, `_physics_process()`, and `get_action_space()` with the following implementation:**

```gdscript

func get_reward() -> float:

	return reward



func _process(_delta: float) -> void:

	# In case of human control, we get the user input

	if control_mode == ControlModes.HUMAN:

		get_user_input()



func _physics_process(_delta: float) -> void:

	# Reset on timeout, this is implemented in parent class to set needs_reset to true,

	# we are re-implementing here to call player.game_over() that handles the game reset.

	n_steps += 1

	if n_steps > reset_after:

		player.game_over(0)

		player.print_game_status("Episode timed out.")



## Defines the actions for the AI agent

func get_action_space() -> Dictionary:

	return {

		"movement": {"size": 5, "action_type": "discrete"},

	}

```

- `get_reward()` just returns the current reward value, this is used by the Godot RL Agents `sync` node to send the reward the data to the training algorithm,
- `_process()` handles user input if human control mode is selected,
- `_physics_process()` handles restarting the game on episode time out,
- `get_action_space()` defines the actions we need to receive from the RL agent to control the robot. In this case that is a discrete action with size 5 (allowing the robot to move in any of 4 directions, or stand still).

**Replace `set_action()`, and `get_user_input()` with the following implementation:**



```gdscript

## Applies AI control actions to the robot

func set_action(action = null) -> void:

	# We have specified discrete action type with size 5,

	# which means there are 5 possible values that the agent can output

	# for each step, i.e. one of: 0, 1, 2, 3, 4,

	# we use those to allow the agent to move in 4 directions,

	# + there is a 'no movement' action.

	# First convert to int to use match as action value is of float type.

	match int(action.movement):

		0:

			player.requested_movement = Vector3.LEFT

		1:

			player.requested_movement = Vector3.RIGHT

		2:

			player.requested_movement = Vector3.FORWARD

		3:

			player.requested_movement = Vector3.BACK

		4:

			player.requested_movement = Vector3.ZERO



## Applies user input actions to the robot

func get_user_input() -> void:

	if Input.is_action_just_pressed("move_up"):

		player.requested_movement = Vector3.FORWARD

	elif Input.is_action_just_pressed("move_right"):

		player.requested_movement = Vector3.RIGHT

	elif Input.is_action_just_pressed("move_down"):

		player.requested_movement = Vector3.BACK

	elif Input.is_action_just_pressed("move_left"):

		player.requested_movement = Vector3.LEFT

```



- `set_action()` converts the action received from the RL agent to a movement direction for the robot.

- `get_user_input()` converts user input into a movement direction for the robot (if human control mode is selected).



Now the implementation is done and we’re ready to export the game and start training!



### Export the game for training:



You can export the game from Godot using `Project > Export`.



# Training:



### **Download a copy of the [SB3 Example script](https://github.com/edbeeching/godot_rl_agents/blob/main/examples/stable_baselines3_example.py) from the Godot RL Repository.**

### **Run training using the arguments below:**

```gdscript

stable_baselines3_example.py --timesteps=250_000 --onnx_export_path=model.onnx --env_path="PathToExportedExecutable" --n_parallel=4 --speedup=32

```

**Set the env path to the exported game and start training.**

Training time may take a while, depending on the hardware and settings. 

The `ep_rew_mean` after the 250 000 training steps should be `1.5+` for the preconfigured map layout.

In case you’d like to tweak the settings, you could change `n_parallel` to get more FPS on some computers, note this could affect the number of timesteps needed to train successfully as well. If you’re familiar with Stable Baselines 3, you can also adjust the [PPO hyperparameters](https://stable-baselines3.readthedocs.io/en/master/modules/ppo.html) by modifying the Python script.

### **Copy the onnx file or the path to the exported onnx file.**

After training completes, you will get a message in the console such as `Exporting onnx to: ...`. You can copy or move the `.onnx` file from there to the Godot project folder, or you can just copy the full path to the file.

### Open the `onnx_inference_scene.tscn` scene:

> [!TIP]
> You can search for “onnx_inference” in the FileSystem search.



Click on the Sync node, then paste the path to the onnx model file from the previous step into the `Onnx Model Path` property. You can adjust the `speed_up` optionally to make the env run faster, recommended for preview is the `0.1` default.

![onnx_scene](https://github.com/user-attachments/assets/38c25d91-6846-45de-9411-728212482628)

### Start the game and watch the agent play:

With the `onnx_inference_scene` open, press `F6` to start the game from this scene. 

> [!TIP]
> You can also open `robot.tscn`, click on the `Robot` node, and enable `Print Game Status Enabled` in inspector to see the debug info printed in the console (whether an episode was successful or not), then get back to the `onnx_inference_scene` before pressing `F6`.

The trained agent should behave similarly as in the video below:

https://github.com/user-attachments/assets/82858efb-1af0-41c5-a211-0d5a85665bf0

If so, 🎉 congratulations! 🌟 You’ve completed the tutorial successfully!

If there are issues with the performance of the agent, you could try retraining with some more time-steps. 

If there are some errors when trying to run the onnx file, check if there is an error message. You could also try importing and running the `Completed` Godot project to see if it works well, then compare for any differences.

# Customizing the map layout:

If you’d like to change the layout of the map, open the `scripts/grid_map.gd` script (you can find it in the FileSystem). 

You will find the relevant code section in the `set_cells()` method:

```gdscript

## You can set the layout by adjusting each row with set_row_cells()

## Note: Changing size after initial start is not supported

## you can change the order or rows or how many of the second tiles to add

## as long as the total size (number of rows, width) remains the same



func set_cells():

	remove_all_tiles()

	

	add_row(Tile.TileNames.orange, Tile.TileNames.goal, 1)

	add_row(Tile.TileNames.orange)

	add_row(Tile.TileNames.orange, Tile.TileNames.tree, 2)

	add_row(Tile.TileNames.road)

	add_row(Tile.TileNames.orange, Tile.TileNames.tree, 2)

	add_row(Tile.TileNames.orange)

	add_row(Tile.TileNames.orange)

	

	set_player_position_to_last_row()



	tiles_instantiated = true

```

In the code above, each `add_row()` method call adds another row to the map. The method takes 3 arguments:

- The name of the tile to use for the first/main tile of that row
- The name of the tile to use for the second tile of that row (the second tile gets randomly positioned along grid x coordinate)
- How many of the second tiles to place in that row.

For instance, a call to:

```gdscript

add_row(Tile.TileNames.orange, Tile.TileNames.tree, 2)

```

will add a grid row with mostly orange (walkable) tiles, and 2 randomly positioned trees (along the x axis).

Note that for any road segment that you add using:

```gdscript

add_row(Tile.TileNames.road)

```

a car will also be added automatically that will move along the road. When using road tiles, you shouldn’t add a second tile as this wasn’t intended by the current design. 

Here’s a more complex layout example you can try:

```gdscript

## You can set the layout by adjusting each row with set_row_cells()

## Note: Changing size after initial start is not supported

## you can change the order or rows or how many of the second tiles to add

## as long as the total size (number of rows, width) remains the same



func set_cells():

	remove_all_tiles()

	

	add_row(Tile.TileNames.orange, Tile.TileNames.goal, 1)

	add_row(Tile.TileNames.orange)

	add_row(Tile.TileNames.orange, Tile.TileNames.tree, 3)

	add_row(Tile.TileNames.road)

	add_row(Tile.TileNames.orange, Tile.TileNames.tree, 3)

	add_row(Tile.TileNames.road)

	add_row(Tile.TileNames.orange, Tile.TileNames.tree, 3)

	add_row(Tile.TileNames.road)

	add_row(Tile.TileNames.orange)

	add_row(Tile.TileNames.orange)

	

	set_player_position_to_last_row()



	tiles_instantiated = true

```

That would generate maps that looks similar to below:

![larger_map](https://github.com/user-attachments/assets/b638f4e8-d2f2-4ee4-a18c-5bac7f8b3aec)

If you’d like to shuffle the rows to create a more randomized map layout, you can try something similar to below:

```gdscript

## You can set the layout by adjusting each row with set_row_cells()

## Note: Changing size after initial start is not supported

## you can change the order or rows or how many of the second tiles to add

## as long as the total size (number of rows, width) remains the same



func set_cells():

	remove_all_tiles()

	

	# Keeps the goal row fixed

	set_row_tiles(0, Tile.TileNames.orange, Tile.TileNames.goal, 1)

	

	var row_ids = range(1, 5)

	row_ids.shuffle()

	set_row_tiles(row_ids[0], Tile.TileNames.orange, Tile.TileNames.tree, 2)

	set_row_tiles(row_ids[1], Tile.TileNames.orange, Tile.TileNames.tree, 2)

	set_row_tiles(row_ids[2], Tile.TileNames.road)

	set_row_tiles(row_ids[3], Tile.TileNames.road)

	

	# Keeps the player starting row fixed

	set_row_tiles(5, Tile.TileNames.orange)

	grid_size_z = 6

	

	set_player_position_to_last_row()



	tiles_instantiated = true

```

Note that these layouts can take longer to train. For example, I tried training the “shuffle” variant above with `n_steps=32,batch_size=256` set in the Python training script, `n_parallel=6` set as CL argument, and I've also increased the observation settings so that the agent can see more rows in front of the player, and one row behind. I’ve trained it for a couple of hours on my PC (I stopped manually, it’s possible the behavior was learned earlier). It resulted in a high success rate during onnx inference.

Some notes:

- During training, you could randomize/shuffle the row order, the number of cars (automatically set based on number of rows with road) can also change, but changing the map size (number of rows or grid width) between episodes is not supported (it’s possible to extend the script to support this case).
- While you can run inference using the same onnx you previously trained on a different map, the results may not be good (might be a bit better if training a shuffled variant to learn a more generalized behavior). Retraining might be needed.
- To see a different map layout properly, you may need to adjust the camera position in Godot editor (found in `res://scenes/game_scene.tscn`).
- Training may take longer for more complex maps. Tuning the Python script hyperparameters might help.
- When the map is larger and takes longer to complete, you may need to increase the timeout steps, this is adjustable from Godot editor as the `reset after` property of AIController3D (check screenshot below):

![reset_after](https://github.com/user-attachments/assets/bf09bb10-c4b0-4942-a04c-c2ea59c3632c)

# Conclusion:

Congrats! You’ve learned how to train a RL agent to successfully complete a grid-like env using Godot, Godot RL Agents, and Stable Baselines 3.

For more inspiration, feel free to check out our other [example environments](https://github.com/edbeeching/godot_rl_agents_examples/tree/main/examples), and our two tutorials on the HF Deep RL Course (https://huggingface.co/learn/deep-rl-course/unitbonus3/godotrl, https://huggingface.co/learn/deep-rl-course/unitbonus5/introduction).