- Deep Reinforcement Learning Hands-On
- Maxim Lapan
- 179字
- 2021-06-25 20:46:58
Chapter 6. Deep Q-Networks
In the previous chapter, we became familiar with the Bellman equation and the practical method of its application called Value iteration. This approach allowed us to significantly improve our speed and convergence in the FrozenLake environment, which is promising, but can we go further?
In this chapter, we'll try to apply the same theory to problems of much greater complexity: arcade games from the Atari 2600 platform, which are the de-facto benchmark of the RL research community. To deal with this new and more challenging goal, we'll talk about problems with the Value iteration method and introduce its variation, called Q-learning. In particular, we'll look at the application of Q-learning to so-called "grid world" environments, which is called tabular Q-learning, and then we'll discuss Q-learning in conjunction with neural networks. This combination has the name DQN. At the end of the chapter, we'll reimplement a DQN algorithm from the famous paper, Playing Atari with Deep Reinforcement Learning by V. Mnih and others, published in 2013, which started a new era in RL development.
- 電氣自動化專業(yè)英語(第3版)
- Oracle SOA Governance 11g Implementation
- 空間機器人遙操作系統(tǒng)及控制
- Cloud Analytics with Microsoft Azure
- 計算機網(wǎng)絡技術實訓
- 水晶石精粹:3ds max & ZBrush三維數(shù)字靜幀藝術
- Spark大數(shù)據(jù)技術與應用
- Splunk Operational Intelligence Cookbook
- PVCBOT機器人控制技術入門
- Docker on Amazon Web Services
- 分析力!專業(yè)Excel的制作與分析實用法則
- Extending Ansible
- 未來學徒:讀懂人工智能飛馳時代
- 菜鳥起飛電腦組裝·維護與故障排查
- 工廠電氣控制設備