Gibbs sampler as a Markov process

AI Thread Summary
The Gibbs sampler is indeed a Markov process, as its transition probabilities depend only on the current state, represented by the vector of values at time t. Each updated variable x_i(t+1) is contingent on the entire set of current variables, but this does not violate the Markov property. The key point is that the transition probabilities can be expressed as functions of the current state, allowing for dependencies on past states without disqualifying it as a Markov process. Many authoritative sources emphasize that transition probabilities can depend on past states, as long as they are conditioned through the current state. Understanding this clarifies the relationship between the Gibbs sampler and Markov chains.
Stochasticus
Messages
2
Reaction score
0
I'm trying to learn more about Markov chains and came across the Gibbs sampler

x_1{t+1} ~ p(x_1|x_2 = x_2{t},...x_n{t})
x_2{t+1} ~ p(x_2|x_1 = x_1{t+1},x_3 = x_3{t},...,x_n{t})
.
.
.
x_i{t+1} ~ p(x_i|x_1 = x_1{t+1},...,x_(i-1) = x_(i-1){t+1},x_(i+1) = x_(i+1){t},...,x_n{t})

Supposedly this thing is a Markov chain. I just don't see it. It sure looks like each updated variable x_i{t+1} is contingent on the whole set, not just the prior value x_i{t}. Can someone show me how this satisfies the Markov criterion
 
Physics news on Phys.org
A process can be viewed as a Markov process if the transition probabilities to the next "state" depends only on the current "state". You can define a state in any way that you wish and it might even consist of an infinite number of variables. Until it is revealed how a state is defined, it is impossible to say whether a given mathematical or physical process is Markov.

The state in the Gibbs sampler is the current vector of values (x_1(t),x_2(t),...x_n(t)). The fact that these variables at time t are realized by a transition probability that depends on the vector at t-1 doesn't contradict that fact that the transisition probabilities are determined by the vector of variables at t. The fact that you might be able to write the transition probabilities at time t as a function of the x_i() at an earlier time such as t = 2 doesn't contradict the fact that the process is a Markov process. It's a Markov process as long as you can expresses the transition probabilities as a function of the current "state".
 
Hey Stochasticus and welcome to the forums.

Yours is an interesting question as I have just assumed this in my prior courses without bothering to check it myself, but I found a PDF that shows a proof that might help you.

http://users.aims.ac.za/~ioana/notes-ch4-2x1.pdf

Scroll down to proposition 4.5
 
That proof doesn't bother to prove the Gibbs Sampler is a Markov Chain. The proof just assumes it is, which I think is justified.

A further thought: Many respectable books phrase their definition of a Markov chain to say that the transistion probabilities to the next state "depend on the history of past states only through the current state". They say it that way instead of saying that the transitions probabilities "don't depend on the history of past states", which is the phrase that casual writers use. The casual writers are wrong. It is actually OK for the transition probabilities to depend on the history of past states and, in fact, you can usually write the transition probabilities as functions of a history of past states.
 
It seems I was overthinking this. I was aware that for the whole set of variables, call it big X = {x_{1},...,x_{n}} the sampler met the Markov criterion. I got caught up trying to prove that each subsequent value x_{i}^{t+1} was conditioned only on x_{i}^{t}.
 
I'm taking a look at intuitionistic propositional logic (IPL). Basically it exclude Double Negation Elimination (DNE) from the set of axiom schemas replacing it with Ex falso quodlibet: ⊥ → p for any proposition p (including both atomic and composite propositions). In IPL, for instance, the Law of Excluded Middle (LEM) p ∨ ¬p is no longer a theorem. My question: aside from the logic formal perspective, is IPL supposed to model/address some specific "kind of world" ? Thanks.
I was reading a Bachelor thesis on Peano Arithmetic (PA). PA has the following axioms (not including the induction schema): $$\begin{align} & (A1) ~~~~ \forall x \neg (x + 1 = 0) \nonumber \\ & (A2) ~~~~ \forall xy (x + 1 =y + 1 \to x = y) \nonumber \\ & (A3) ~~~~ \forall x (x + 0 = x) \nonumber \\ & (A4) ~~~~ \forall xy (x + (y +1) = (x + y ) + 1) \nonumber \\ & (A5) ~~~~ \forall x (x \cdot 0 = 0) \nonumber \\ & (A6) ~~~~ \forall xy (x \cdot (y + 1) = (x \cdot y) + x) \nonumber...
Back
Top