官术网_书友最值得收藏!

Algorithms covered in this book

In Chapter 2, Temporal Difference, SARSA, and Q-Learning, we will look into our first two RL algorithms: Q-learning and SARSA. Both of these algorithms are tabular-based and do not require the use of neural networks. Thus, we will code them in Python and NumPy. In Chapter 3, Deep Q-Network, we will cover DQN and use TensorFlow to code the agent for the rest of the book. We will then train it to play Atari Breakout. In Chapter 4, Double DQN, Dueling Architectures, and Rainbow, we will cover double DQN, dueling network architectures, and rainbow DQN. In Chapter 5, Deep Deterministic Policy Gradient, we will look at our first Actor-Critic RL algorithm called DDPG, learn about policy gradients, and apply them to a continuous action problem. In Chapter 6, Asynchronous Methods – A3C and A2C, we will investigate A3C, which is another RL algorithm that uses a master and several worker processes. In Chapter 7, Trust Region Policy Optimization and Proximal Policy Optimization, we will investigate two more RL algorithms: TRPO and PPO. Finally, we will apply DDPG and PPO to train an agent to drive a car autonomously in Chapter 8, Deep RL Applied to Autonomous Driving. From Chapter 3, Deep Q-Network, to Chapter 8, Deep RL Applied to Autonomous Driving, we'll use TensorFlow agents. Have fun learning RL.

主站蜘蛛池模板: 定日县| 铜梁县| 宜都市| 大新县| 涿州市| 左权县| 红河县| 柘荣县| 清徐县| 茶陵县| 梨树县| 梓潼县| 太仆寺旗| 嘉禾县| 深泽县| 建水县| 鞍山市| 武定县| 乌兰察布市| 西昌市| 武宁县| 墨竹工卡县| 清新县| 阜康市| 南汇区| 禄丰县| 兴海县| 育儿| 镇沅| 阿尔山市| 额济纳旗| 西乌珠穆沁旗| 玉林市| 高雄市| 方正县| 淮安市| 利辛县| 永泰县| 林西县| 扎赉特旗| 应用必备|