Summary
RL is one of the fundamental paradigms under the umbrella of machine learning. The principles of RL are very general and interdisciplinary, and they are not bound to a specific application.
RL considers the interaction of an agent with an external environment, taking inspiration from the human learning process. RL explicitly targets the need to explore efficiently and the exploration-exploitation trade-off appearing in almost all human problems; this is a peculiarity that distinguishes this discipline from others.
We started this chapter with a high-level description of RL, showing some interesting applications. We then introduced the main concepts of RL, describing what an agent is, what an environment is, and how an agent interacts with its environment. Finally, we implemented Gym and Baselines by showing how these libraries make RL extremely simple.
In the next chapter, we will learn more about the theory behind RL, starting with Markov chains and arriving at MDPs. We will present the two functions at the core of almost all RL algorithms, namely the state-value function, which evaluates the goodness of states, and the action-value function, which evaluates the quality of the state-action pair.
- Arduino入門基礎(chǔ)教程
- Learning SQL Server Reporting Services 2012
- 用“芯”探核:龍芯派開發(fā)實戰(zhàn)
- 觸摸屏實用技術(shù)與工程應(yīng)用
- FPGA從入門到精通(實戰(zhàn)篇)
- 現(xiàn)代辦公設(shè)備使用與維護
- 深入淺出SSD:固態(tài)存儲核心技術(shù)、原理與實戰(zhàn)(第2版)
- 平衡掌控者:游戲數(shù)值經(jīng)濟設(shè)計
- Artificial Intelligence Business:How you can profit from AI
- 筆記本電腦維修不是事兒(第2版)
- Hands-On Machine Learning with C#
- Mastering Adobe Photoshop Elements
- Building 3D Models with modo 701
- 基于Proteus仿真的51單片機應(yīng)用
- 微控制器的應(yīng)用