• No results found

Markov Chains (273023), Exercise session 8, Tue 12 March 2013.

N/A
N/A
Protected

Academic year: 2021

Share "Markov Chains (273023), Exercise session 8, Tue 12 March 2013."

Copied!
2
0
0

Loading.... (view fulltext now)

Full text

(1)

Markov Chains (273023), Exercise session 8, Tue 12 March 2013. This exercise session does not give any extra points to the exam.

The exercises in this session cover the whole course and can be con- sidered as representative of typical questions in the exam. There are three questions in the exam Friday 15 March.

Exercise 8.1. Decide (guess) whether the following statements are true or false. Each correct answer is worth 1 point, wrong answers decrease the total by 1 point each.

(1) The row sums and the column sums equal one for every tran- sition probability matrix of a Markov chain with finite state space.

(2) There exists a Markov chain with finite state space such that one of the states is aperiodic while another state is periodic.

(3) Let P be the transition probability matrix of a time reversible Markov chain Xt with finite state space Ω. Then the transpo- sition of P defines the time reversal of Xt.

(4) Every aperiodic Markov chain with finite state space has a unique stationary probability distribution.

(5) The simple random walk Xt on the n-cycle can be considered as the projection of the simple random walk Yt on Z with the equivalence relation “is congruent to, modulo n”.

(6) The behvaiour of the simple random walk on the integers can be analyzed using reflection principle.

Exercise 8.2. Decide (guess) whether the following statements are true or false. Each correct answer is worth 1 point, wrong answers decrease the total by 1 point each.

(1) Let µ be a probability distribution on a finite group G. The random walk on G with increment distribution µ is irreducible if

S = {g ∈ G : µ(g) > 0}

generates G.

(2) Let Ω be finite and let X and Y be two random variables defined on Ω. Define µ(x) = P(X = x) and ν(y) = P(Y = y). Then (X, Y ) is a coupling of µ and ν.

(3) Let P be the transition probability matrix of an irreducible Markov chain with finite state space Ω with at least two ele- ments. Then zero is an eigenvalue of P .

(4) Let Ω be finite and π an arbitrary probability distribution on Ω. Then there exists an irreducible, aperiodic Markov chain Xt with stationary probability π.

1

(2)

2

(5) Every aperiodic, irreducible Markov chain with countable state space has a stationary probability distribution.

(6) An irreducible Markov chain with finite state space is recurrent.

Exercise 8.3. Let P be an irreducible transition matrix of period b > 0.

Show carefully that the finite state space Ω can be partitioned into b sets C1, . . . , Cb ⊂ Ω in such a way that P (x, y) > 0 if and only if x ∈ Ci

and y ∈ Ci+1 for some i with the addition i + 1 modulo b.

Exercise 8.4. Show carefully that the uniform distribution is stationary for a random walk on a group with symmetric increment distribution µ.

Exercise 8.5. Show that there exists a Markov chain Xt with state space Ω = {1, 2, 3} and transition probability matrix P such that µ = (1, 0, 0) and ν = (0, 1/2, 1/2) are stationary for Xt.

Exercise 8.6. Let Xt be a Markov chain with finite state space Ω, and aperiodic and irreducible transition probability matrix P . Let Yt = (Xt+1, Xt) for t = 0, 1, . . . . Show that Yt is an irreducible Markov chain on the state space

Ω := {(x, y) : P (x, y) > 0}.˜

References

Related documents

The results in this paper could still be used to test such restrictions, but then they are identical in form to those in linear VAR:s - even if the drift and the covariance terms

Assuming that the system starts at state W , i.e. Four white and four black balls are distributed in two urns in such a way that each contains four balls. At each step, we draw one

Consider the following process (Glauber dynamics for q-colorings): Select a random vertex of G and change the color of the vertex to any free color that is not take by any of

Suppose that P is a transition probability matrix of an irreducible Markov chain (with possibly countably many states).. Let P be a symmetric transition probability matrix of

Att förhöjningen är störst för parvis Gibbs sampler beror på att man på detta sätt inte får lika bra variation mellan de i tiden närliggande vektorerna som när fler termer

Finally, Subsection 2.3 introduces options on the CDS index, sometimes denoted by credit index options, and uses the result form Subsection 2.2 to provide a formula for the payoff

We showed the efficiency of the MCMC based method and the tail behaviour for various parameters and distributions respectively.. For the tail behaviour, we can conclude that the

Like before, the problem with too few data (see chapter 4.3) appears for the third order Markov chain, it appears for the weekly prediction when using window length 4, 5, 7 and