Gibbs sampler as a Markov process

Click For Summary

Discussion Overview

The discussion centers on the Gibbs sampler and its classification as a Markov process. Participants explore the relationship between the Gibbs sampler's transition probabilities and the Markov criterion, examining whether the sampler meets the necessary conditions for being a Markov chain.

Discussion Character

  • Exploratory
  • Technical explanation
  • Debate/contested

Main Points Raised

  • One participant questions how the Gibbs sampler can be a Markov chain, noting that each updated variable appears contingent on the entire set of variables rather than just its previous value.
  • Another participant explains that a process can be considered Markov if the transition probabilities depend solely on the current state, which they define as the vector of values at time t.
  • A third participant shares a resource that purportedly contains a proof related to the Gibbs sampler's Markov properties, although they express uncertainty about its rigor.
  • A later reply suggests that the proof does not adequately demonstrate the Gibbs sampler's status as a Markov chain, arguing that it is acceptable for transition probabilities to depend on past states as long as they can be expressed in terms of the current state.
  • One participant reflects on their initial overthinking, acknowledging that the Gibbs sampler meets the Markov criterion when considering the entire set of variables rather than individual updates.

Areas of Agreement / Disagreement

Participants express differing views on the Gibbs sampler's classification as a Markov process, with some asserting it meets the criteria while others remain skeptical or seek further clarification. The discussion does not reach a consensus.

Contextual Notes

Some participants highlight the ambiguity in defining the "current state" and the implications of conditioning on past states, which may affect the interpretation of the Gibbs sampler as a Markov process.

Stochasticus
Messages
2
Reaction score
0
I'm trying to learn more about Markov chains and came across the Gibbs sampler

x_1{t+1} ~ p(x_1|x_2 = x_2{t},...x_n{t})
x_2{t+1} ~ p(x_2|x_1 = x_1{t+1},x_3 = x_3{t},...,x_n{t})
.
.
.
x_i{t+1} ~ p(x_i|x_1 = x_1{t+1},...,x_(i-1) = x_(i-1){t+1},x_(i+1) = x_(i+1){t},...,x_n{t})

Supposedly this thing is a Markov chain. I just don't see it. It sure looks like each updated variable x_i{t+1} is contingent on the whole set, not just the prior value x_i{t}. Can someone show me how this satisfies the Markov criterion
 
Physics news on Phys.org
A process can be viewed as a Markov process if the transition probabilities to the next "state" depends only on the current "state". You can define a state in any way that you wish and it might even consist of an infinite number of variables. Until it is revealed how a state is defined, it is impossible to say whether a given mathematical or physical process is Markov.

The state in the Gibbs sampler is the current vector of values (x_1(t),x_2(t),...x_n(t)). The fact that these variables at time t are realized by a transition probability that depends on the vector at t-1 doesn't contradict that fact that the transisition probabilities are determined by the vector of variables at t. The fact that you might be able to write the transition probabilities at time t as a function of the x_i() at an earlier time such as t = 2 doesn't contradict the fact that the process is a Markov process. It's a Markov process as long as you can expresses the transition probabilities as a function of the current "state".
 
Hey Stochasticus and welcome to the forums.

Yours is an interesting question as I have just assumed this in my prior courses without bothering to check it myself, but I found a PDF that shows a proof that might help you.

http://users.aims.ac.za/~ioana/notes-ch4-2x1.pdf

Scroll down to proposition 4.5
 
That proof doesn't bother to prove the Gibbs Sampler is a Markov Chain. The proof just assumes it is, which I think is justified.

A further thought: Many respectable books phrase their definition of a Markov chain to say that the transistion probabilities to the next state "depend on the history of past states only through the current state". They say it that way instead of saying that the transitions probabilities "don't depend on the history of past states", which is the phrase that casual writers use. The casual writers are wrong. It is actually OK for the transition probabilities to depend on the history of past states and, in fact, you can usually write the transition probabilities as functions of a history of past states.
 
It seems I was overthinking this. I was aware that for the whole set of variables, call it big X = {x_{1},...,x_{n}} the sampler met the Markov criterion. I got caught up trying to prove that each subsequent value x_{i}^{t+1} was conditioned only on x_{i}^{t}.
 

Similar threads

  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 25 ·
Replies
25
Views
4K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 7 ·
Replies
7
Views
2K
  • · Replies 21 ·
Replies
21
Views
2K
  • · Replies 2 ·
Replies
2
Views
3K