官术网_书友最值得收藏!

Algorithms covered in this book

In Chapter 2, Temporal Difference, SARSA, and Q-Learning, we will look into our first two RL algorithms: Q-learning and SARSA. Both of these algorithms are tabular-based and do not require the use of neural networks. Thus, we will code them in Python and NumPy. In Chapter 3, Deep Q-Network, we will cover DQN and use TensorFlow to code the agent for the rest of the book. We will then train it to play Atari Breakout. In Chapter 4, Double DQN, Dueling Architectures, and Rainbow, we will cover double DQN, dueling network architectures, and rainbow DQN. In Chapter 5, Deep Deterministic Policy Gradient, we will look at our first Actor-Critic RL algorithm called DDPG, learn about policy gradients, and apply them to a continuous action problem. In Chapter 6, Asynchronous Methods – A3C and A2C, we will investigate A3C, which is another RL algorithm that uses a master and several worker processes. In Chapter 7, Trust Region Policy Optimization and Proximal Policy Optimization, we will investigate two more RL algorithms: TRPO and PPO. Finally, we will apply DDPG and PPO to train an agent to drive a car autonomously in Chapter 8, Deep RL Applied to Autonomous Driving. From Chapter 3, Deep Q-Network, to Chapter 8, Deep RL Applied to Autonomous Driving, we'll use TensorFlow agents. Have fun learning RL.

主站蜘蛛池模板: 武冈市| 徐水县| 潍坊市| 定日县| 葵青区| 平武县| 柘城县| 沧州市| 阳春市| 庐江县| 调兵山市| 嘉义市| 万年县| 晋宁县| 巴林左旗| 宁河县| 桃源县| 泽普县| 大洼县| 盱眙县| 安顺市| 石门县| 瑞丽市| 道孚县| 老河口市| 登封市| 鄂伦春自治旗| 登封市| 凤翔县| 东方市| 宝坻区| 宁国市| 大埔县| 房山区| 阳谷县| 哈密市| 宜川县| 东兴市| 确山县| 衡阳县| 绍兴市|