- PyTorch 1.x Reinforcement Learning Cookbook
- Yuxi (Hayden) Liu
- 249字
- 2021-06-24 12:34:39
How it works...
Using Gym, we can easily create an environment instance by calling the make() method with the name of the environment as the parameter.
As you may have noticed, the actions that the agent performs are randomly chosen using the sample() method.
Note that, normally, we would have a more sophisticated agent guided by reinforcement learning algorithms. Here, we just demonstrated how to simulate an environment, and how an agent takes actions regardless of the outcome.
Run this a few times and see what we get:
>>> env.action_space.sample()
0
>>> env.action_space.sample()
3
>>> env.action_space.sample()
0
>>> env.action_space.sample()
4
>>> env.action_space.sample()
2
>>> env.action_space.sample()
1
>>> env.action_space.sample()
4
>>> env.action_space.sample()
5
>>> env.action_space.sample()
1
>>> env.action_space.sample()
0
There are six possible actions in total. We can also see this by running the following command:
>>> env.action_space
Discrete(6)
Actions from 0 to 5 stand for No Operation, Fire, Up, Right, Left, and Down, respectively, which are all the moves the spaceship in the game can do.
The step() method will let the agent take the action that is specified as its parameter. The render() method will update the display window based on the latest observation of the environment.
The observation of the environment, new_state, is represented by a 210 x 160 x 3 matrix, as follows:
>>> print(new_state.shape)
(210, 160, 3)
This means that each frame of the display screen is an RGB image of size 210 x 160.
- Hands-On Deep Learning with Apache Spark
- Ansible Configuration Management
- 影視后期編輯與合成
- 工業機器人運動仿真編程實踐:基于Android和OpenGL
- 分數階系統分析與控制研究
- 電子設備及系統人機工程設計(第2版)
- ESP8266 Robotics Projects
- Hands-On Dashboard Development with QlikView
- 網絡服務器搭建與管理
- Kubernetes on AWS
- Flash CS3動畫制作
- 51單片機應用開發實戰手冊
- Architecting Cloud Computing Solutions
- Office 2010辦公應用
- 筆記本電腦維修實用教程