官术网_书友最值得收藏!

How it works...

Using Gym, we can easily create an environment instance by calling the make() method with the name of the environment as the parameter.

As you may have noticed, the actions that the agent performs are randomly chosen using the sample() method.

Note that, normally, we would have a more sophisticated agent guided by reinforcement learning algorithms. Here, we just demonstrated how to simulate an environment, and how an agent takes actions regardless of the outcome.

Run this a few times and see what we get:

>>> env.action_space.sample()
0
>>> env.action_space.sample()
3
>>> env.action_space.sample()
0
>>> env.action_space.sample()
4
>>> env.action_space.sample()
2
>>> env.action_space.sample()
1
>>> env.action_space.sample()
4
>>> env.action_space.sample()
5
>>> env.action_space.sample()
1
>>> env.action_space.sample()
0

There are six possible actions in total. We can also see this by running the following command:

>>> env.action_space
Discrete(6)

Actions from 0 to 5 stand for No Operation, Fire, Up, Right, Left, and Down, respectively, which are all the moves the spaceship in the game can do.

The step() method will let the agent take the action that is specified as its parameter. The render() method will update the display window based on the latest observation of the environment.

The observation of the environment, new_state, is represented by a 210 x 160 x 3 matrix, as follows:

>>> print(new_state.shape)
(210, 160, 3)

This means that each frame of the display screen is an RGB image of size 210 x 160.

主站蜘蛛池模板: 霍州市| 定结县| 丰镇市| 德保县| 怀柔区| 青海省| 宜宾市| 淅川县| 嘉荫县| 壶关县| 福贡县| 浠水县| 塔城市| 西畴县| 清苑县| 平阴县| 农安县| 岱山县| 新密市| 湘乡市| 永丰县| 无棣县| 平潭县| 岳阳市| 绩溪县| 阿合奇县| 都安| 五寨县| 特克斯县| 隆化县| 海原县| 资中县| 鹿泉市| 武冈市| 应用必备| 郧西县| 北辰区| 祁东县| 高阳县| 会泽县| 定远县|