Proving P(X_4|X_1,X_2)=P(X_4|X_2) for Markov Chain with Finite Possibilities

In summary: For discrete states, it would seem that the proof is necessary.To cover the case of continuous states, replace the probabilities by probability density functions and the sums by integrals over the range ##C_3##.To cover the case of continuous time, replace times 1, 2, 3, 4 by times ##0\le t_1\le t_2\le t_3\le t_4##,This seems to be what I was looking for. For continuous states I think a proof is unnecessary, since the definition requires it. For discrete states, it would seem that the proof is necessary.
  • #1
mathman
Science Advisor
8,140
572
TL;DR Summary
Requirement for Markov chain.
Given events ##X_i## and the following: ##P(X_3|X_1,X_2)=P(X_3|X_2)## and ##P(X_4|X_1,X_2,X_3)=P(X_4|X_3)## prove ##P(X_4|X_1,X_2)=P(X_4|X_2)##?

Special case proof: Finite number of possibilities for each ##X_i##. Transition matrix from ##X_2## to ##X_4## is product of transitions from ##X_2## to ##X_3## to ##X_4##. Generalize?
 
Last edited:
Physics news on Phys.org
  • #2
Are you sure this is true? It doesn't feel true. We can partition ##X=X_1\cup X_2\cup X_3\cup X_4## into 15 disjoint subsets ( ##=2^4 - 1##), and the only relationship between their probabilities is that they must be non-negative and sum to ##P(X)##. When we bring in the first two equations above, that still doesn't look like enough constraints on the 15 unknown probabilities to give us the third formula.

Given that, my approach would be to try to construct a counter-example by choosing values for the 15 disjoint components. If we find that we the third formula holds no matter how we change the 15 probabilities, the numbers should give us a clue as to why the relationship holds.

Note this does not use the properties of Markov Chains at all. I can't see how they would relate to the problem as stated.
 
  • #3
andrewkirk said:
Are you sure this is true? It doesn't feel true. We can partition ##X=X_1\cup X_2\cup X_3\cup X_4## into 15 disjoint subsets ( ##=2^4 - 1##), and the only relationship between their probabilities is that they must be non-negative and sum to ##P(X)##. When we bring in the first two equations above, that still doesn't look like enough constraints on the 15 unknown probabilities to give us the third formula.

Given that, my approach would be to try to construct a counter-example by choosing values for the 15 disjoint components. If we find that we the third formula holds no matter how we change the 15 probabilities, the numbers should give us a clue as to why the relationship holds.

Note this does not use the properties of Markov Chains at all. I can't see how they would relate to the problem as stated.
The givens are the bare bones requirements for a sequence to be a Markov chain.
I couldn't follow you. Could you give the example?
 
  • #4
mathman said:
The givens are the bare bones requirements for a sequence to be a Markov chain.
A Markov chain is defined as a process in which the probability of future events, as measured at time t, depends only on the state at time t. Markov chains can have continuous or discrete states and they can progress over continuous or discrete time. Finite state processes have finite transition matrices whereas continuous state processes have transition distributions.

How does a formula about four undefined events ##X_1,X_2,X_3,X_4##, with no context, have anything to do with that?
 
  • #5
andrewkirk said:
A Markov chain is defined as a process in which the probability of future events, as measured at time t, depends only on the state at time t. Markov chains can have continuous or discrete states and they can progress over continuous or discrete time. Finite state processes have finite transition matrices whereas continuous state processes have transition distributions.

How does a formula about four undefined events ##X_1,X_2,X_3,X_4##, with no context, have anything to do with that?
Consider these to be the first four states in a Markov chain. The statements given are the third state depends only on the second, independent of the first, while the fourth state depends only on the third, independent of the first two.

The question is:, ignore the third state. Is the fourth state dependent only on the second, independent of the first?
 
  • #6
Proof:

Let ##C_t## be the set of possible states at time ##t##.

\begin{align*}
P(X_4|X_1=x_1,X_2=x_2)
&=
\sum_{x_3\in C_3}
P(X_4|X_1=x_1,X_2=x_2, X_3=x_3)
P(X_3=x_3|X_1=x_1,X_2=x_2)
\\
&=
\sum_{x_3\in C_3}
P(X_4|X_3=x_3)
P(X_3=x_3|X_1=x_1,X_2=x_2)
\\ &\quad\quad\textrm{by first given equation}
\\
&=
\sum_{x_3\in C_3}
P(X_4|X_3=x_3)
P(X_3=x_3|X_2=x_2)
\\ &\quad\quad\textrm{by second given equation}
\\ &=
P(X_4=x_4|X_2=x_2)
\end{align*}

To cover the case of continuous states, replace the probabilities by probability density functions and the sums by integrals over the range ##C_3##.

To cover the case of continuous time, replace times 1, 2, 3, 4 by times ##0\le t_1\le t_2\le t_3\le t_4##,
 
  • #7
This seems to be what I was looking for. For continuous states I think a proof is unnecessary, since the definition requires it.
 

1. What is a Markov Chain with Finite Possibilities?

A Markov Chain with Finite Possibilities is a mathematical model used to describe the probabilistic behavior of a system over time. It is characterized by a set of states and transition probabilities between those states.

2. How does a Markov Chain with Finite Possibilities work?

A Markov Chain with Finite Possibilities works by assuming that the probability of transitioning from one state to another only depends on the current state, and not on any previous states. This is known as the Markov property.

3. What does the equation P(X_4|X_1,X_2)=P(X_4|X_2) mean in the context of a Markov Chain with Finite Possibilities?

This equation represents the conditional probability of being in state X_4 at time 4, given that the system was in states X_1 and X_2 at times 1 and 2, respectively. It is equal to the conditional probability of being in state X_4 at time 4, given that the system was in state X_2 at time 2. This is a key property of Markov Chains, known as the Markov property.

4. Why is it important to prove P(X_4|X_1,X_2)=P(X_4|X_2) for a Markov Chain with Finite Possibilities?

Proving this equation is important because it verifies that the Markov Chain is indeed a valid model for the system being studied. It also allows for the use of various mathematical techniques and tools to analyze the behavior and make predictions about the system.

5. What are some real-world applications of Markov Chains with Finite Possibilities?

Markov Chains with Finite Possibilities have a wide range of applications, including modeling stock market trends, predicting weather patterns, and analyzing the behavior of complex systems such as biological networks and social networks. They are also commonly used in machine learning and artificial intelligence algorithms.

Similar threads

Replies
3
Views
737
  • Introductory Physics Homework Help
Replies
4
Views
741
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
2K
  • Precalculus Mathematics Homework Help
Replies
32
Views
846
  • Set Theory, Logic, Probability, Statistics
Replies
2
Views
4K
  • Special and General Relativity
Replies
2
Views
674
  • Set Theory, Logic, Probability, Statistics
Replies
1
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
2K
Back
Top