官术网_书友最值得收藏!

Creating a Markov chain

Let's get started by creating a Markov chain, on which the MDP is developed.

A Markov chain describes a sequence of events that comply with the Markov property. It is defined by a set of possible states, S = {s0, s1, ... , sm}, and a transition matrix, T(s, s'), consisting of the probabilities of state s transitioning to state s'. With the Markov property, the future state of the process, given the present state, is conditionally independent of past states. In other words, the state of the process at t+1 is dependent only on the state at t. Here, we use a process of study and sleep as an example and create a Markov chain based on two states, s0 (study) and s1 (sleep). Let's say we have the following transition matrix:

In the next section, we will compute the transition matrix after k steps, and the probabilities of being in each state given an initial distribution of states, such as [0.7, 0.3], meaning there is a 70% chance that the process starts with study and a 30% chance that it starts with sleep.

主站蜘蛛池模板: 南丹县| 永胜县| 绥滨县| 江西省| 阳信县| 红桥区| 达州市| 鹤壁市| 饶阳县| 巴东县| 蒙阴县| 历史| 卓尼县| 湛江市| 昭苏县| 汪清县| 若尔盖县| 轮台县| 阳谷县| 独山县| 兴和县| 若尔盖县| 长阳| 虹口区| 漯河市| 建始县| 济宁市| 烟台市| 永修县| 南安市| 博兴县| 湖州市| 西吉县| 左贡县| 开阳县| 钟祥市| 阿城市| 刚察县| 彝良县| 浠水县| 成安县|