Summary
In this chapter, we learned what the Markov chain and Markov process are and how RL problems are represented using MDP. We have also looked at the Bellman equation, and we solved the Bellman equation to derive an optimal policy using DP. In the Chapter 4, Gaming with Monte Carlo Methods, we will look at the Monte Carlo tree search and how to build intelligent games using it.
推薦閱讀
- 漫話大數(shù)據(jù)
- 數(shù)據(jù)挖掘原理與實踐
- ETL數(shù)據(jù)整合與處理(Kettle)
- 達夢數(shù)據(jù)庫編程指南
- Creating Mobile Apps with Sencha Touch 2
- 工業(yè)大數(shù)據(jù)分析算法實戰(zhàn)
- Dependency Injection with AngularJS
- 數(shù)字媒體交互設計(初級):Web產(chǎn)品交互設計方法與案例
- 數(shù)據(jù)庫設計與應用(SQL Server 2014)(第二版)
- INSTANT Android Fragmentation Management How-to
- 數(shù)據(jù)庫應用系統(tǒng)開發(fā)實例
- Unity 2018 By Example(Second Edition)
- 大數(shù)據(jù)分析:R基礎及應用
- Microsoft Dynamics NAV 2015 Professional Reporting
- 從Lucene到Elasticsearch:全文檢索實戰(zhàn)