- PyTorch 1.x Reinforcement Learning Cookbook
- Yuxi (Hayden) Liu
- 76字
- 2021-06-24 12:34:44
There's more...
In fact, irrespective of the initial state the process was in, the state distribution will always converge to [0.5714, 0.4286]. You could test with other initial distributions, such as [0.2, 0.8] and [1, 0]. The distribution will remain [0.5714, 0.4286] after 10 steps.
A Markov chain does not necessarily converge, especially when it contains transient or current states. But if it does converge, it will reach the same equilibrium regardless of the starting distribution.
推薦閱讀
- Hands-On Internet of Things with MQTT
- Visualforce Development Cookbook(Second Edition)
- Deep Learning Quick Reference
- 計算機網絡應用基礎
- 流處理器研究與設計
- Java Web整合開發全程指南
- INSTANT Autodesk Revit 2013 Customization with .NET How-to
- Implementing Splunk 7(Third Edition)
- 人工智能趣味入門:光環板程序設計
- Excel 2007技巧大全
- Windows Server 2003系統安全管理
- 過程控制系統
- 大數據案例精析
- Mastering MongoDB 3.x
- 一步步寫嵌入式操作系統